Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Applying GPGPU to Recurrent Neural Network Language Model based Fast Network Search in the Real-Time LVCSR (2007.11794v1)

Published 23 Jul 2020 in cs.CL and cs.LG

Abstract: Recurrent Neural Network LLMs (RNNLMs) have started to be used in various fields of speech recognition due to their outstanding performance. However, the high computational complexity of RNNLMs has been a hurdle in applying the RNNLM to a real-time Large Vocabulary Continuous Speech Recognition (LVCSR). In order to accelerate the speed of RNNLM-based network searches during decoding, we apply the General Purpose Graphic Processing Units (GPGPUs). This paper proposes a novel method of applying GPGPUs to RNNLM-based graph traversals. We have achieved our goal by reducing redundant computations on CPUs and amount of transfer between GPGPUs and CPUs. The proposed approach was evaluated on both WSJ corpus and in-house data. Experiments shows that the proposed approach achieves the real-time speed in various circumstances while maintaining the Word Error Rate (WER) to be relatively 10% lower than that of n-gram models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Kyungmin Lee (52 papers)
  2. Chiyoun Park (4 papers)
  3. Ilhwan Kim (4 papers)
  4. Namhoon Kim (8 papers)
  5. Jaewon Lee (39 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.