Domain Adaptation for Arabic Machine Translation: The Case of Financial Texts (2309.12863v1)
Abstract: Neural machine translation (NMT) has shown impressive performance when trained on large-scale corpora. However, generic NMT systems have demonstrated poor performance on out-of-domain translation. To mitigate this issue, several domain adaptation methods have recently been proposed which often lead to better translation quality than genetic NMT systems. While there has been some continuous progress in NMT for English and other European languages, domain adaption in Arabic has received little attention in the literature. The current study, therefore, aims to explore the effectiveness of domain-specific adaptation for Arabic MT (AMT), in yet unexplored domain, financial news articles. To this end, we developed carefully a parallel corpus for Arabic-English (AR- EN) translation in the financial domain for benchmarking different domain adaptation methods. We then fine-tuned several pre-trained NMT and LLMs including ChatGPT-3.5 Turbo on our dataset. The results showed that the fine-tuning is successful using just a few well-aligned in-domain AR-EN segments. The quality of ChatGPT translation was superior than other models based on automatic and human evaluations. To the best of our knowledge, this is the first work on fine-tuning ChatGPT towards financial domain transfer learning. To contribute to research in domain translation, we made our datasets and fine-tuned models available at https://huggingface.co/asas-ai/.
- “Six challenges for neural machine translation” In arXiv preprint arXiv:1706.03872, 2017
- Hal Daumé Iii and Jagadeesh Jagarlamudi “Domain adaptation for machine translation by mining unseen words” In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, 2011, pp. 407–412
- Danielle Saunders “Domain adaptation and multi-domain adaptation for neural machine translation: A survey” In Journal of Artificial Intelligence Research 75, 2022, pp. 351–424
- “A survey of domain adaptation for machine translation” In Journal of information processing 28 Information Processing Society of Japan, 2020, pp. 413–426
- Yasmin Moslem, Rejwanul Haque and Andy Way “Adaptive machine translation with large language models” In arXiv preprint arXiv:2301.13294, 2023
- “Transforming machine translation: a deep learning system reaches news translation quality comparable to human professionals” In Nature communications 11.1 Nature Publishing Group, 2020, pp. 1–15
- “Syntactic reordering for Arabic-English phrase-based machine translation” In Database theory and application, bio-science and bio-technology Springer, 2010, pp. 198–206
- Zakaryia Mustafa Almahasees “Assessment of Google and Microsoft Bing translation of journalistic texts” In International Journal of Languages, Literature and Linguistics 4.3, 2018, pp. 231–235
- Rico Sennrich, Barry Haddow and Alexandra Birch “Neural machine translation of rare words with subword units” In arXiv preprint arXiv:1508.07909, 2015
- Marta R Costa-Jussa and José AR Fonollosa “Character-based neural machine translation” In arXiv preprint arXiv:1603.00810, 2016
- Minh-Thang Luong and Christopher D Manning “Achieving open vocabulary neural machine translation with hybrid word-character models” In arXiv preprint arXiv:1604.00788, 2016
- Mathias Müller, Annette Rios and Rico Sennrich “Domain robustness in neural machine translation” In arXiv preprint arXiv:1911.03109, 2019
- Mai Oudah, Amjad Almahairi and Nizar Habash “The impact of preprocessing on Arabic-English statistical and neural machine translation” In arXiv preprint arXiv:1906.11751, 2019
- “Domain-Specific Text Generation for Machine Translation” In arXiv preprint arXiv:2208.05909, 2022
- “BLEU: a method for automatic evaluation of machine translation” In Proceedings of the 40th annual meeting on association for computational linguistics, 2002, pp. 311–318 Association for Computational Linguistics
- Ankur Bapna, Naveen Arivazhagan and Orhan Firat “Simple, scalable adaptation for neural machine translation” In arXiv preprint arXiv:1909.08478, 2019
- Shuoheng Yang, Yuxin Wang and Xiaowen Chu “A survey of deep learning techniques for neural machine translation” In arXiv preprint arXiv:2002.07526, 2020
- Ilya Sutskever, Oriol Vinyals and Quoc V Le “Sequence to sequence learning with neural networks” In Advances in neural information processing systems 27, 2014
- “Google’s multilingual neural machine translation system: Enabling zero-shot translation” In Transactions of the Association for Computational Linguistics 5 MIT Press, 2017, pp. 339–351
- Dzmitry Bahdanau, Kyunghyun Cho and Yoshua Bengio “Neural machine translation by jointly learning to align and translate” In arXiv preprint arXiv:1409.0473, 2014
- “Deep recurrent models with fast-forward connections for neural machine translation” In Transactions of the Association for Computational Linguistics 4 MIT Press, 2016, pp. 371–383
- Duygu Ataman, Wilker Aziz and Alexandra Birch “A latent morphology model for open-vocabulary neural machine translation” In arXiv preprint arXiv:1910.13890, 2019
- “Encoding source language with convolutional neural network for machine translation” In arXiv preprint arXiv:1503.01838, 2015
- “A convolutional encoder model for neural machine translation” In arXiv preprint arXiv:1611.02344, 2016
- “Neural machine translation in linear time” In arXiv preprint arXiv:1610.10099, 2016
- “Convolutional sequence to sequence learning” In International conference on machine learning, 2017, pp. 1243–1252 PMLR
- “Attention is all you need” In Advances in neural information processing systems 30, 2017
- Minh-Thang Luong, Hieu Pham and Christopher D Manning “Effective approaches to attention-based neural machine translation” In arXiv preprint arXiv:1508.04025, 2015
- Felix Stahlberg “Neural machine translation: A review” In Journal of Artificial Intelligence Research 69, 2020, pp. 343–418
- “Bert: Pre-training of deep bidirectional transformers for language understanding” In arXiv preprint arXiv:1810.04805, 2018
- Wissam Antoun, Fady Baly and Hazem Hajj “Arabert: Transformer-based model for Arabic language understanding” In arXiv preprint arXiv:2003.00104, 2020
- “Gigabert: Zero-shot transfer learning from english to arabic” In Proceedings of The 2020 Conference on Empirical Methods on Natural Language Processing (EMNLP), 2020
- “Exploring transfer learning and domain data selection for the biomedical translation” In Proceedings of the Fourth Conference on Machine Translation (Volume 3: Shared Task Papers, Day 2), 2019, pp. 156–163
- “ParaMed: a parallel corpus for English–Chinese translation in the biomedical domain” In BMC Medical Informatics and Decision Making 21.1 Springer, 2021, pp. 258
- “Customized Neural Machine Translation Systems for the Swiss Legal Domain” In Proceedings of the 14th Conference of the Association for Machine Translation in the Americas (Volume 2: User Track), 2020, pp. 217–223
- “Post-editing productivity with neural machine translation: An empirical assessment of speed and quality in the banking and finance domain” In arXiv preprint arXiv:1906.01685, 2019
- “The trade-offs of domain adaptation for neural language models” In arXiv preprint arXiv:2109.10274, 2021
- “Investigating Backtranslation in Neural Machine Translation. 2018” Arxiv, 1804
- Chenhui Chu, Raj Dabre and Sadao Kurohashi “An empirical comparison of domain adaptation methods for neural machine translation” In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2017, pp. 385–391
- “Factorized transformer for multi-domain neural machine translation” In Findings of the Association for Computational Linguistics: EMNLP 2020, 2020, pp. 4221–4230
- Hitham Abo Bakr, Khaled Shaalan and Ibrahim Ziedan “A hybrid approach for converting written Egyptian colloquial dialect into diacritized Arabic” In The 6th international conference on informatics and systems, infos2008. cairo university, 2008 Citeseer
- Emad Mohamed, Behrang Mohit and Kemal Oflazer “Transforming standard Arabic to colloquial Arabic” In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2012, pp. 176–180
- “Improving Arabic-Chinese statistical machine translation using English as pivot language” In Proceedings of the Fourth Workshop on Statistical Machine Translation, 2009, pp. 173–181
- El Moatez Billah Nagoudi, AbdelRahim Elmadany and Muhammad Abdul-Mageed “AraT5: Text-to-text transformers for Arabic language generation” In arXiv preprint arXiv:2109.12068, 2021
- “Arabic Machine Translation: A Survey With Challenges and Future Directions” In IEEE Access 9 IEEE, 2021, pp. 161445–161468
- Mohamed Seghir Hadj Ameur, Farid Meziane and Ahmed Guessoum “Arabic machine translation: A survey of the latest trends and challenges” In Computer Science Review 38 Elsevier, 2020, pp. 100305
- Nizar Y Habash “Introduction to Arabic natural language processing” In Synthesis lectures on human language technologies 3.1 Morgan & Claypool Publishers, 2010, pp. 1–187
- Rana Ehab, Eslam Amer and Mahmoud Gadallah “English-Arabic hybrid machine translation system using EBMT and translation memory” In Int. J. Adv. Comput. Sci. Appl. 10.1, 2019, pp. 195–203
- Wissam Antoun, Fady Baly and Hazem Hajj “AraGPT2: Pre-Trained Transformer for Arabic Language Generation” In Proceedings of the Sixth Arabic Natural Language Processing Workshop Kyiv, Ukraine (Virtual): Association for Computational Linguistics, 2021, pp. 196–207 URL: https://aclanthology.org/2021.wanlp-1.21
- “Language Models are Unsupervised Multitask Learners”, 2019
- “GPTAraEval: A Comprehensive Evaluation of ChatGPT on Arabic NLP”, 2023 arXiv:2305.14976 [cs.CL]
- “OPUS-MT – Building open translation services for the World” In European Association for Machine Translation Conferences/Workshops, 2020 URL: https://api.semanticscholar.org/CorpusID:221097277
- Jörg Tiedemann “The Tatoeba Translation Challenge – Realistic Data Sets for Low Resource and Multilingual MT” In Proceedings of the Fifth Conference on Machine Translation Online: Association for Computational Linguistics, 2020, pp. 1174–1182 URL: https://www.aclweb.org/anthology/2020.wmt-1.139
- “SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing” In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018: System Demonstrations, Brussels, Belgium, October 31 - November 4, 2018, 2018, pp. 66–71 DOI: 10.18653/v1/d18-2012
- “No Language Left Behind: Scaling Human-Centered Machine Translation”, 2022 arXiv:2207.04672 [cs.CL]
- Maja Popović “chrF: character n-gram F-score for automatic MT evaluation” In Proceedings of the Tenth Workshop on Statistical Machine Translation Lisbon, Portugal: Association for Computational Linguistics, 2015, pp. 392–395 DOI: 10.18653/v1/W15-3049
- “A study of translation edit rate with targeted human annotation” In Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers, 2006, pp. 223–231
- “COMET: A Neural Framework for MT Evaluation” In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) Online: Association for Computational Linguistics, 2020, pp. 2685–2702 DOI: 10.18653/v1/2020.emnlp-main.213
- “BERTScore: Evaluating Text Generation with BERT”, 2020 arXiv:1904.09675 [cs.CL]
- “Towards Making the Most of ChatGPT for Machine Translation” In ArXiv abs/2303.13780, 2023 URL: https://api.semanticscholar.org/CorpusID:257704711
- Faten Khoshafah “ChatGPT for Arabic-English Translation: Evaluating the Accuracy” In Research Square Research Square, 2023, pp. https://doi.org/10.21203/rs.3.rs–2814154/v1
- “Taqyim: Evaluating Arabic NLP Tasks Using ChatGPT Models” In arXiv preprint arXiv:2306.16322, 2023
- El Moatez Billah Nagoudi, AbdelRahim Elmadany and Muhammad Abdul-Mageed “AraT5: Text-to-Text Transformers for Arabic Language Generation”, 2022 arXiv:2109.12068 [cs.CL]
- “How Good Are GPT Models at Machine Translation? A Comprehensive Evaluation”, 2023 arXiv:2302.09210 [cs.CL]
- Deborah Coughlin “Correlating automated and human assessments of machine translation quality” In Proceedings of Machine Translation Summit IX: Papers, 2003
- Emad A. Alghamdi (9 papers)
- Jezia Zakraoui (2 papers)
- Fares A. Abanmy (1 paper)