(2015. 6) Teaching Machine Read And Comprehend
published in 2015. 6
Karl Moritz Hermann, Tomáš Kočiský, Edward Grefenstette, Lasse Espeholt, Will Kay, Mustafa Suleyman and Phil Blunsom
Simple Summary
Define a new methodology that resolves the bottleneck and provides large scale supervised reading comprehension data.
Dataset : CNN and Daily Mail
Model
Deep LSTM: 2-layer bidirectional LSTM without attention mechanism
Attentive reader: 1-layer bidirectional LSTM with attention mechanism for the whole query (focus on the passages of a context document that are most likely to inform the answer to the query)
Impatient Reader: 1-layer bidirectional LSTM with attention mechanism for each token in the query (allows the model to recurrently accumulate information
from the document as it sees each query token, ultimately outputting a final joint document query
representation for the answer prediction)
Uniform Reader: Uniform attention to all document tokens
Attention visualization
Last updated