Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Utilizing Lexical Similarity between Related, Low-resource Languages for Pivot-based SMT (1702.07203v2)

Published 23 Feb 2017 in cs.CL

Abstract: We investigate pivot-based translation between related languages in a low resource, phrase-based SMT setting. We show that a subword-level pivot-based SMT model using a related pivot language is substantially better than word and morpheme-level pivot models. It is also highly competitive with the best direct translation model, which is encouraging as no direct source-target training corpus is used. We also show that combining multiple related language pivot models can rival a direct translation model. Thus, the use of subwords as translation units coupled with multiple related pivot languages can compensate for the lack of a direct parallel corpus.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Anoop Kunchukuttan (45 papers)
  2. Maulik Shah (4 papers)
  3. Pradyot Prakash (4 papers)
  4. Pushpak Bhattacharyya (153 papers)
Citations (8)