Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
157 tokens/sec
GPT-4o
8 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Joint Space Neural Probabilistic Language Model for Statistical Machine Translation (1301.3614v3)

Published 16 Jan 2013 in cs.CL

Abstract: A neural probabilistic LLM (NPLM) provides an idea to achieve the better perplexity than n-gram LLM and their smoothed LLMs. This paper investigates application area in bilingual NLP, specifically Statistical Machine Translation (SMT). We focus on the perspectives that NPLM has potential to open the possibility to complement potentially huge' monolingual resources into theresource-constraint' bilingual resources. We introduce an ngram-HMM LLM as NPLM using the non-parametric Bayesian construction. In order to facilitate the application to various tasks, we propose the joint space model of ngram-HMM LLM. We show an experiment of system combination in the area of SMT. One discovery was that our treatment of noise improved the results 0.20 BLEU points if NPLM is trained in relatively small corpus, in our case 500,000 sentence pairs, which is often the case due to the long training time of NPLM.

Citations (2)

Summary

We haven't generated a summary for this paper yet.