Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Analyzing Architectures for Neural Machine Translation Using Low Computational Resources (2111.03813v1)

Published 6 Nov 2021 in cs.CL

Abstract: With the recent developments in the field of Natural Language Processing, there has been a rise in the use of different architectures for Neural Machine Translation. Transformer architectures are used to achieve state-of-the-art accuracy, but they are very computationally expensive to train. Everyone cannot have such setups consisting of high-end GPUs and other resources. We train our models on low computational resources and investigate the results. As expected, transformers outperformed other architectures, but there were some surprising results. Transformers consisting of more encoders and decoders took more time to train but had fewer BLEU scores. LSTM performed well in the experiment and took comparatively less time to train than transformers, making it suitable to use in situations having time constraints.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Aditya Mandke (4 papers)
  2. Onkar Litake (11 papers)
  3. Dipali Kadam (6 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.