Papers
Topics
Authors
Recent
Search
2000 character limit reached

Design of an Open-Source Architecture for Neural Machine Translation

Published 6 Mar 2024 in cs.CL and cs.AI | (2403.03582v1)

Abstract: adaptNMT is an open-source application that offers a streamlined approach to the development and deployment of Recurrent Neural Networks and Transformer models. This application is built upon the widely-adopted OpenNMT ecosystem, and is particularly useful for new entrants to the field, as it simplifies the setup of the development environment and creation of train, validation, and test splits. The application offers a graphing feature that illustrates the progress of model training, and employs SentencePiece for creating subword segmentation models. Furthermore, the application provides an intuitive user interface that facilitates hyperparameter customization. Notably, a single-click model development approach has been implemented, and models developed by adaptNMT can be evaluated using a range of metrics. To encourage eco-friendly research, adaptNMT incorporates a green report that flags the power consumption and kgCO${_2}$ emissions generated during model development. The application is freely available.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (31)
  1. 2020. Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible ai. Information Fusion, 58:82–115.
  2. 2019. Massively multilingual sentence embeddings for zero-shot cross-lingual transfer and beyond. Transactions of the Association for Computational Linguistics, 7:597–610.
  3. 2021. Evaluating the carbon footprint of NLP methods: a survey and analysis of existing tools. In Proceedings of the Second Workshop on Simple and Efficient Natural Language Processing, pages 11–21, Virtual, November. Association for Computational Linguistics.
  4. 2021. On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, FAccT ’21, page 610–623, New York, NY, USA. Association for Computing Machinery.
  5. 2017. Findings of the 2017 conference on machine translation (WMT17). In Proceedings of the Second Conference on Machine Translation, pages 169–214, Copenhagen, Denmark, September. Association for Computational Linguistics.
  6. 2018. Findings of the 2018 conference on machine translation (WMT18). In Proceedings of the Third Conference on Machine Translation: Shared Task Papers, pages 272–303, Belgium, Brussels, October. Association for Computational Linguistics.
  7. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
  8. 2022. No language left behind: Scaling human-centered machine translation. arXiv preprint arXiv:2207.04672.
  9. 2014. Meteor universal: Language specific translation evaluation for any target language. In Proceedings of the Ninth Workshop on Statistical Machine Translation, pages 376–380, Baltimore, Maryland, USA, June. Association for Computational Linguistics.
  10. 2020. Towards the systematic reporting of the energy and carbon footprints of machine learning. Journal of Machine Learning Research, 21(248):1–43.
  11. 2022a. Knowledge distillation: A method for making neural machine translation more efficient. Information, 13(2).
  12. 2022b. Knowledge distillation for sustainable neural machine translation. In Proceedings of the 15th Biennial Conference of the Association for Machine Translation in the Americas (Volume 2: Users and Providers Track and Government Track), pages 221–230, Orlando, USA, September. Association for Machine Translation in the Americas.
  13. 2018. Marian: Fast neural machine translation in C++. In Proceedings of ACL 2018, System Demonstrations, pages 116–121, Melbourne, Australia, July. Association for Computational Linguistics.
  14. 2017. OpenNMT: Open-source toolkit for neural machine translation. In Proceedings of ACL 2017, System Demonstrations, pages 67–72, Vancouver, Canada, July. Association for Computational Linguistics.
  15. 2019. Quantifying the carbon emissions of machine learning. arXiv preprint arXiv:1910.09700.
  16. 2021a. Transformers for low-resource languages: Is féidir linn! In Proceedings of Machine Translation Summit XVIII: Research Track, pages 48–60.
  17. 2021b. Machine translation in the covid domain: an English-Irish case study for LoResMT 2021. In Proceedings of the 4th Workshop on Technologies for MT of Low Resource Languages (LoResMT2021), pages 144–150, Virtual, August. Association for Machine Translation in the Americas.
  18. 2022a. gaHealth: An English–Irish bilingual corpus of health data. In Proceedings of the Thirteenth Language Resources and Evaluation Conference, pages 6753–6758, Marseille, France, June. European Language Resources Association.
  19. 2022b. Human evaluation of English–Irish Transformer-Based NMT. Information, 13(7):309. 19pp.
  20. 2023a. adaptNMT: an open-source, language-agnostic development environment for neural machine translation. Language Resources and Evaluation.
  21. 2023b. adaptLLM: Fine-tuning large language models for improved machine translation. an open-source approach with NLLB and GPT-J integration. In Journal of Artificial Intelligence Research [manuscript submitted], El Segundo, CA, United States. AI Access Foundation.
  22. 2023c. Fine-tuning LLMs using low-resource language pairs: A comparative study with loresmt2021 shared task baselines. In Proceedings of Machine Translation Summit XIX: Research Track [manuscript submitted], Macau SAR, China, September 4-8. Asia-Pacific Association for Machine Translation (AAMT).
  23. 2003. Precision and recall of machine translation. In Companion Volume of the Proceedings of HLT-NAACL 2003 - Short Papers, pages 61–63, Edmonton, Canada.
  24. 2019. fairseq: A fast, extensible toolkit for sequence modeling. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (Demonstrations), pages 48–53, Minneapolis, Minnesota, June. Association for Computational Linguistics.
  25. 2002. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311–318, Philadelphia, PA.
  26. Popović, Maja. 2015. chrF: character n-gram F-score for automatic MT evaluation. In Proceedings of the Tenth Workshop on Statistical Machine Translation, pages 392–395, Lisbon, Portugal, September. Association for Computational Linguistics.
  27. Post, Matt. 2018. A call for clarity in reporting BLEU scores. In Proceedings of the Third Conference on Machine Translation: Research Papers, pages 186–191, Brussels, Belgium, October. Association for Computational Linguistics.
  28. 2006. A study of translation edit rate with targeted human annotation. In Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers, pages 223–231, Cambridge, Massachusetts, USA, August 8-12. Association for Machine Translation in the Americas.
  29. 2017. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, page 6000–6010, Red Hook, NY, USA. Curran Associates Inc.
  30. Way, Andy. 2018. Quality expectations of machine translation. In Moorkens, Joss, Sheila Castilho, Federico Gaspari, and Stephen Doherty, editors, Translation Quality Assessment: From Principles to Practice, pages 159–178. Springer, Cham, Switzerland.
  31. Way, Andy. 2019. Machine translation: Where are we at today? In Angelone, Eric, Gary Massey, and Maureen Ehrensberger-Dow, editors, The Bloomsbury Companion to Language Industry Studies, page 311—332. Bloomsbury, London.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.