2000 character limit reached
Iterative Alternating Neural Attention for Machine Reading (1606.02245v4)
Published 7 Jun 2016 in cs.CL and cs.NE
Abstract: We propose a novel neural attention architecture to tackle machine comprehension tasks, such as answering Cloze-style queries with respect to a document. Unlike previous models, we do not collapse the query into a single vector, instead we deploy an iterative alternating attention mechanism that allows a fine-grained exploration of both the query and the document. Our model outperforms state-of-the-art baselines in standard machine comprehension benchmarks such as CNN news articles and the Children's Book Test (CBT) dataset.
- Alessandro Sordoni (53 papers)
- Philip Bachman (25 papers)
- Adam Trischler (50 papers)
- Yoshua Bengio (601 papers)