Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

English Broadcast News Speech Recognition by Humans and Machines (1904.13258v1)

Published 30 Apr 2019 in cs.CL, cs.SD, and eess.AS

Abstract: With recent advances in deep learning, considerable attention has been given to achieving automatic speech recognition performance close to human performance on tasks like conversational telephone speech (CTS) recognition. In this paper we evaluate the usefulness of these proposed techniques on broadcast news (BN), a similar challenging task. We also perform a set of recognition measurements to understand how close the achieved automatic speech recognition results are to human performance on this task. On two publicly available BN test sets, DEV04F and RT04, our speech recognition system using LSTM and residual network based acoustic models with a combination of n-gram and neural network LLMs performs at 6.5% and 5.9% word error rate. By achieving new performance milestones on these test sets, our experiments show that techniques developed on other related tasks, like CTS, can be transferred to achieve similar performance. In contrast, the best measured human recognition performance on these test sets is much lower, at 3.6% and 2.8% respectively, indicating that there is still room for new techniques and improvements in this space, to reach human performance levels.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Samuel Thomas (42 papers)
  2. Masayuki Suzuki (6 papers)
  3. Yinghui Huang (13 papers)
  4. Gakuto Kurata (13 papers)
  5. Zoltan Tuske (14 papers)
  6. George Saon (39 papers)
  7. Brian Kingsbury (54 papers)
  8. Michael Picheny (32 papers)
  9. Tom Dibert (1 paper)
  10. Alice Kaiser-Schatzlein (2 papers)
  11. Bern Samko (1 paper)
Citations (14)

Summary

We haven't generated a summary for this paper yet.