TY - JOUR
T1 - Long-term memory networks for question answering
AU - Ma, Fenglong
AU - Chitta, Radha
AU - Kataria, Saurabh
AU - Zhou, Jing
AU - Ramesh, Palghat
AU - Sun, Tong
AU - Gao, Jing
N1 - Publisher Copyright:
Copyright © by the paper's authors. Copying permitted for private and academic purposes.
PY - 2017
Y1 - 2017
N2 - Question answering is an important and difficult task in the natural language process-ing domain, because many basic natural lan-guage processing tasks can be cast into a ques-tion answering task. Several deep neural net-work architectures have been developed re-cently, which employ memory and inference components to memorize and reason over text information, and generate answers to ques-tions. However, a major drawback of many such models is that they are capable of only generating single-word answers. In addition, they require large amount of training data to generate accurate answers. In this paper, we introduce the Long-Term Memory Network (LTMN), which incorporates both an exter-nal memory module and a Long Short-Term Memory (LSTM) module to comprehend the input data and generate multi-word answers. The LTMN model can be trained end-to-end using back-propagation and requires minimal supervision. We test our model on two syn-thetic data sets (based on Facebook's bAbI data set) and the real-world Stanford ques-tion answering data set, and show that it can achieve state-of-the-art performance.
AB - Question answering is an important and difficult task in the natural language process-ing domain, because many basic natural lan-guage processing tasks can be cast into a ques-tion answering task. Several deep neural net-work architectures have been developed re-cently, which employ memory and inference components to memorize and reason over text information, and generate answers to ques-tions. However, a major drawback of many such models is that they are capable of only generating single-word answers. In addition, they require large amount of training data to generate accurate answers. In this paper, we introduce the Long-Term Memory Network (LTMN), which incorporates both an exter-nal memory module and a Long Short-Term Memory (LSTM) module to comprehend the input data and generate multi-word answers. The LTMN model can be trained end-to-end using back-propagation and requires minimal supervision. We test our model on two syn-thetic data sets (based on Facebook's bAbI data set) and the real-world Stanford ques-tion answering data set, and show that it can achieve state-of-the-art performance.
UR - http://www.scopus.com/inward/record.url?scp=85034946837&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85034946837&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85034946837
SN - 1613-0073
VL - 1986
SP - 7
EP - 14
JO - CEUR Workshop Proceedings
JF - CEUR Workshop Proceedings
T2 - 2017 IJCAI Workshop on Semantic Machine Learning, SML 2017
Y2 - 20 August 2017
ER -