2000 character limit reached
On the long-term learning ability of LSTM LMs (2106.08927v1)
Published 16 Jun 2021 in cs.CL and cs.LG
Abstract: We inspect the long-term learning ability of Long Short-Term Memory LLMs (LSTM LMs) by evaluating a contextual extension based on the Continuous Bag-of-Words (CBOW) model for both sentence- and discourse-level LSTM LMs and by analyzing its performance. We evaluate on text and speech. Sentence-level models using the long-term contextual module perform comparably to vanilla discourse-level LSTM LMs. On the other hand, the extension does not provide gains for discourse-level models. These findings indicate that discourse-level LSTM LMs already rely on contextual information to perform long-term learning.
- Wim Boes (8 papers)
- Robbe Van Rompaey (2 papers)
- Lyan Verwimp (11 papers)
- Joris Pelemans (7 papers)
- Hugo Van hamme (59 papers)
- Patrick Wambacq (5 papers)