Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language Processing (2006.07116v1)

Published 12 Jun 2020 in cs.LG, cs.CL, and stat.ML

Abstract: Neural Architecture Search (NAS) is a promising and rapidly evolving research area. Training a large number of neural networks requires an exceptional amount of computational power, which makes NAS unreachable for those researchers who have limited or no access to high-performance clusters and supercomputers. A few benchmarks with precomputed neural architectures performances have been recently introduced to overcome this problem and ensure more reproducible experiments. However, these benchmarks are only for the computer vision domain and, thus, are built from the image datasets and convolution-derived architectures. In this work, we step outside the computer vision domain by leveraging the LLMing task, which is the core of NLP. Our main contribution is as follows: we have provided search space of recurrent neural networks on the text datasets and trained 14k architectures within it; we have conducted both intrinsic and extrinsic evaluation of the trained models using datasets for semantic relatedness and language understanding evaluation; finally, we have tested several NAS algorithms to demonstrate how the precomputed results can be utilized. We believe that our results have high potential of usage for both NAS and NLP communities.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Nikita Klyuchnikov (10 papers)
  2. Ilya Trofimov (16 papers)
  3. Ekaterina Artemova (53 papers)
  4. Mikhail Salnikov (11 papers)
  5. Maxim Fedorov (3 papers)
  6. Evgeny Burnaev (189 papers)
Citations (96)

Summary

We haven't generated a summary for this paper yet.