Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Modality Attention for End-to-End Audio-visual Speech Recognition (1811.05250v2)

Published 13 Nov 2018 in cs.CL, cs.CV, cs.SD, and eess.AS

Abstract: Audio-visual speech recognition (AVSR) system is thought to be one of the most promising solutions for robust speech recognition, especially in noisy environment. In this paper, we propose a novel multimodal attention based method for audio-visual speech recognition which could automatically learn the fused representation from both modalities based on their importance. Our method is realized using state-of-the-art sequence-to-sequence (Seq2seq) architectures. Experimental results show that relative improvements from 2% up to 36% over the auditory modality alone are obtained depending on the different signal-to-noise-ratio (SNR). Compared to the traditional feature concatenation methods, our proposed approach can achieve better recognition performance under both clean and noisy conditions. We believe modality attention based end-to-end method can be easily generalized to other multimodal tasks with correlated information.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Pan Zhou (220 papers)
  2. Wenwen Yang (4 papers)
  3. Wei Chen (1288 papers)
  4. Yanfeng Wang (211 papers)
  5. Jia Jia (59 papers)
Citations (66)