Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Order-Preserving Abstractive Summarization for Spoken Content Based on Connectionist Temporal Classification (1709.05475v2)

Published 16 Sep 2017 in cs.CL

Abstract: Connectionist temporal classification (CTC) is a powerful approach for sequence-to-sequence learning, and has been popularly used in speech recognition. The central ideas of CTC include adding a label "blank" during training. With this mechanism, CTC eliminates the need of segment alignment, and hence has been applied to various sequence-to-sequence learning problems. In this work, we applied CTC to abstractive summarization for spoken content. The "blank" in this case implies the corresponding input data are less important or noisy; thus it can be ignored. This approach was shown to outperform the existing methods in term of ROUGE scores over Chinese Gigaword and MATBN corpora. This approach also has the nice property that the ordering of words or characters in the input documents can be better preserved in the generated summaries.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Bo-Ru Lu (8 papers)
  2. Frank Shyu (1 paper)
  3. Yun-Nung Chen (104 papers)
  4. Hung-yi Lee (327 papers)
  5. Lin-Shan Lee (42 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.