Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Domain Adaptation for Arabic Machine Translation: The Case of Financial Texts (2309.12863v1)

Published 22 Sep 2023 in cs.CL and cs.AI

Abstract: Neural machine translation (NMT) has shown impressive performance when trained on large-scale corpora. However, generic NMT systems have demonstrated poor performance on out-of-domain translation. To mitigate this issue, several domain adaptation methods have recently been proposed which often lead to better translation quality than genetic NMT systems. While there has been some continuous progress in NMT for English and other European languages, domain adaption in Arabic has received little attention in the literature. The current study, therefore, aims to explore the effectiveness of domain-specific adaptation for Arabic MT (AMT), in yet unexplored domain, financial news articles. To this end, we developed carefully a parallel corpus for Arabic-English (AR- EN) translation in the financial domain for benchmarking different domain adaptation methods. We then fine-tuned several pre-trained NMT and LLMs including ChatGPT-3.5 Turbo on our dataset. The results showed that the fine-tuning is successful using just a few well-aligned in-domain AR-EN segments. The quality of ChatGPT translation was superior than other models based on automatic and human evaluations. To the best of our knowledge, this is the first work on fine-tuning ChatGPT towards financial domain transfer learning. To contribute to research in domain translation, we made our datasets and fine-tuned models available at https://huggingface.co/asas-ai/.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (65)
  1. “Six challenges for neural machine translation” In arXiv preprint arXiv:1706.03872, 2017
  2. Hal Daumé Iii and Jagadeesh Jagarlamudi “Domain adaptation for machine translation by mining unseen words” In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, 2011, pp. 407–412
  3. Danielle Saunders “Domain adaptation and multi-domain adaptation for neural machine translation: A survey” In Journal of Artificial Intelligence Research 75, 2022, pp. 351–424
  4. “A survey of domain adaptation for machine translation” In Journal of information processing 28 Information Processing Society of Japan, 2020, pp. 413–426
  5. Yasmin Moslem, Rejwanul Haque and Andy Way “Adaptive machine translation with large language models” In arXiv preprint arXiv:2301.13294, 2023
  6. “Transforming machine translation: a deep learning system reaches news translation quality comparable to human professionals” In Nature communications 11.1 Nature Publishing Group, 2020, pp. 1–15
  7. “Syntactic reordering for Arabic-English phrase-based machine translation” In Database theory and application, bio-science and bio-technology Springer, 2010, pp. 198–206
  8. Zakaryia Mustafa Almahasees “Assessment of Google and Microsoft Bing translation of journalistic texts” In International Journal of Languages, Literature and Linguistics 4.3, 2018, pp. 231–235
  9. Rico Sennrich, Barry Haddow and Alexandra Birch “Neural machine translation of rare words with subword units” In arXiv preprint arXiv:1508.07909, 2015
  10. Marta R Costa-Jussa and José AR Fonollosa “Character-based neural machine translation” In arXiv preprint arXiv:1603.00810, 2016
  11. Minh-Thang Luong and Christopher D Manning “Achieving open vocabulary neural machine translation with hybrid word-character models” In arXiv preprint arXiv:1604.00788, 2016
  12. Mathias Müller, Annette Rios and Rico Sennrich “Domain robustness in neural machine translation” In arXiv preprint arXiv:1911.03109, 2019
  13. Mai Oudah, Amjad Almahairi and Nizar Habash “The impact of preprocessing on Arabic-English statistical and neural machine translation” In arXiv preprint arXiv:1906.11751, 2019
  14. “Domain-Specific Text Generation for Machine Translation” In arXiv preprint arXiv:2208.05909, 2022
  15. “BLEU: a method for automatic evaluation of machine translation” In Proceedings of the 40th annual meeting on association for computational linguistics, 2002, pp. 311–318 Association for Computational Linguistics
  16. Ankur Bapna, Naveen Arivazhagan and Orhan Firat “Simple, scalable adaptation for neural machine translation” In arXiv preprint arXiv:1909.08478, 2019
  17. Shuoheng Yang, Yuxin Wang and Xiaowen Chu “A survey of deep learning techniques for neural machine translation” In arXiv preprint arXiv:2002.07526, 2020
  18. Ilya Sutskever, Oriol Vinyals and Quoc V Le “Sequence to sequence learning with neural networks” In Advances in neural information processing systems 27, 2014
  19. “Google’s multilingual neural machine translation system: Enabling zero-shot translation” In Transactions of the Association for Computational Linguistics 5 MIT Press, 2017, pp. 339–351
  20. Dzmitry Bahdanau, Kyunghyun Cho and Yoshua Bengio “Neural machine translation by jointly learning to align and translate” In arXiv preprint arXiv:1409.0473, 2014
  21. “Deep recurrent models with fast-forward connections for neural machine translation” In Transactions of the Association for Computational Linguistics 4 MIT Press, 2016, pp. 371–383
  22. Duygu Ataman, Wilker Aziz and Alexandra Birch “A latent morphology model for open-vocabulary neural machine translation” In arXiv preprint arXiv:1910.13890, 2019
  23. “Encoding source language with convolutional neural network for machine translation” In arXiv preprint arXiv:1503.01838, 2015
  24. “A convolutional encoder model for neural machine translation” In arXiv preprint arXiv:1611.02344, 2016
  25. “Neural machine translation in linear time” In arXiv preprint arXiv:1610.10099, 2016
  26. “Convolutional sequence to sequence learning” In International conference on machine learning, 2017, pp. 1243–1252 PMLR
  27. “Attention is all you need” In Advances in neural information processing systems 30, 2017
  28. Minh-Thang Luong, Hieu Pham and Christopher D Manning “Effective approaches to attention-based neural machine translation” In arXiv preprint arXiv:1508.04025, 2015
  29. Felix Stahlberg “Neural machine translation: A review” In Journal of Artificial Intelligence Research 69, 2020, pp. 343–418
  30. “Bert: Pre-training of deep bidirectional transformers for language understanding” In arXiv preprint arXiv:1810.04805, 2018
  31. Wissam Antoun, Fady Baly and Hazem Hajj “Arabert: Transformer-based model for Arabic language understanding” In arXiv preprint arXiv:2003.00104, 2020
  32. “Gigabert: Zero-shot transfer learning from english to arabic” In Proceedings of The 2020 Conference on Empirical Methods on Natural Language Processing (EMNLP), 2020
  33. “Exploring transfer learning and domain data selection for the biomedical translation” In Proceedings of the Fourth Conference on Machine Translation (Volume 3: Shared Task Papers, Day 2), 2019, pp. 156–163
  34. “ParaMed: a parallel corpus for English–Chinese translation in the biomedical domain” In BMC Medical Informatics and Decision Making 21.1 Springer, 2021, pp. 258
  35. “Customized Neural Machine Translation Systems for the Swiss Legal Domain” In Proceedings of the 14th Conference of the Association for Machine Translation in the Americas (Volume 2: User Track), 2020, pp. 217–223
  36. “Post-editing productivity with neural machine translation: An empirical assessment of speed and quality in the banking and finance domain” In arXiv preprint arXiv:1906.01685, 2019
  37. “The trade-offs of domain adaptation for neural language models” In arXiv preprint arXiv:2109.10274, 2021
  38. “Investigating Backtranslation in Neural Machine Translation. 2018” Arxiv, 1804
  39. Chenhui Chu, Raj Dabre and Sadao Kurohashi “An empirical comparison of domain adaptation methods for neural machine translation” In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2017, pp. 385–391
  40. “Factorized transformer for multi-domain neural machine translation” In Findings of the Association for Computational Linguistics: EMNLP 2020, 2020, pp. 4221–4230
  41. Hitham Abo Bakr, Khaled Shaalan and Ibrahim Ziedan “A hybrid approach for converting written Egyptian colloquial dialect into diacritized Arabic” In The 6th international conference on informatics and systems, infos2008. cairo university, 2008 Citeseer
  42. Emad Mohamed, Behrang Mohit and Kemal Oflazer “Transforming standard Arabic to colloquial Arabic” In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2012, pp. 176–180
  43. “Improving Arabic-Chinese statistical machine translation using English as pivot language” In Proceedings of the Fourth Workshop on Statistical Machine Translation, 2009, pp. 173–181
  44. El Moatez Billah Nagoudi, AbdelRahim Elmadany and Muhammad Abdul-Mageed “AraT5: Text-to-text transformers for Arabic language generation” In arXiv preprint arXiv:2109.12068, 2021
  45. “Arabic Machine Translation: A Survey With Challenges and Future Directions” In IEEE Access 9 IEEE, 2021, pp. 161445–161468
  46. Mohamed Seghir Hadj Ameur, Farid Meziane and Ahmed Guessoum “Arabic machine translation: A survey of the latest trends and challenges” In Computer Science Review 38 Elsevier, 2020, pp. 100305
  47. Nizar Y Habash “Introduction to Arabic natural language processing” In Synthesis lectures on human language technologies 3.1 Morgan & Claypool Publishers, 2010, pp. 1–187
  48. Rana Ehab, Eslam Amer and Mahmoud Gadallah “English-Arabic hybrid machine translation system using EBMT and translation memory” In Int. J. Adv. Comput. Sci. Appl. 10.1, 2019, pp. 195–203
  49. Wissam Antoun, Fady Baly and Hazem Hajj “AraGPT2: Pre-Trained Transformer for Arabic Language Generation” In Proceedings of the Sixth Arabic Natural Language Processing Workshop Kyiv, Ukraine (Virtual): Association for Computational Linguistics, 2021, pp. 196–207 URL: https://aclanthology.org/2021.wanlp-1.21
  50. “Language Models are Unsupervised Multitask Learners”, 2019
  51. “GPTAraEval: A Comprehensive Evaluation of ChatGPT on Arabic NLP”, 2023 arXiv:2305.14976 [cs.CL]
  52. “OPUS-MT – Building open translation services for the World” In European Association for Machine Translation Conferences/Workshops, 2020 URL: https://api.semanticscholar.org/CorpusID:221097277
  53. Jörg Tiedemann “The Tatoeba Translation Challenge – Realistic Data Sets for Low Resource and Multilingual MT” In Proceedings of the Fifth Conference on Machine Translation Online: Association for Computational Linguistics, 2020, pp. 1174–1182 URL: https://www.aclweb.org/anthology/2020.wmt-1.139
  54. “SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing” In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018: System Demonstrations, Brussels, Belgium, October 31 - November 4, 2018, 2018, pp. 66–71 DOI: 10.18653/v1/d18-2012
  55. “No Language Left Behind: Scaling Human-Centered Machine Translation”, 2022 arXiv:2207.04672 [cs.CL]
  56. Maja Popović “chrF: character n-gram F-score for automatic MT evaluation” In Proceedings of the Tenth Workshop on Statistical Machine Translation Lisbon, Portugal: Association for Computational Linguistics, 2015, pp. 392–395 DOI: 10.18653/v1/W15-3049
  57. “A study of translation edit rate with targeted human annotation” In Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers, 2006, pp. 223–231
  58. “COMET: A Neural Framework for MT Evaluation” In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) Online: Association for Computational Linguistics, 2020, pp. 2685–2702 DOI: 10.18653/v1/2020.emnlp-main.213
  59. “BERTScore: Evaluating Text Generation with BERT”, 2020 arXiv:1904.09675 [cs.CL]
  60. “Towards Making the Most of ChatGPT for Machine Translation” In ArXiv abs/2303.13780, 2023 URL: https://api.semanticscholar.org/CorpusID:257704711
  61. Faten Khoshafah “ChatGPT for Arabic-English Translation: Evaluating the Accuracy” In Research Square Research Square, 2023, pp. https://doi.org/10.21203/rs.3.rs–2814154/v1
  62. “Taqyim: Evaluating Arabic NLP Tasks Using ChatGPT Models” In arXiv preprint arXiv:2306.16322, 2023
  63. El Moatez Billah Nagoudi, AbdelRahim Elmadany and Muhammad Abdul-Mageed “AraT5: Text-to-Text Transformers for Arabic Language Generation”, 2022 arXiv:2109.12068 [cs.CL]
  64. “How Good Are GPT Models at Machine Translation? A Comprehensive Evaluation”, 2023 arXiv:2302.09210 [cs.CL]
  65. Deborah Coughlin “Correlating automated and human assessments of machine translation quality” In Proceedings of Machine Translation Summit IX: Papers, 2003
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Emad A. Alghamdi (9 papers)
  2. Jezia Zakraoui (2 papers)
  3. Fares A. Abanmy (1 paper)
Citations (1)