TY - JOUR
T1 - Joint learning of question answering and question generation
AU - Sun, Yibo
AU - Tang, Duyu
AU - Duan, Nan
AU - Qin, Tao
AU - Liu, Shujie
AU - Yan, Zhao
AU - Zhou, Ming
AU - Lv, Yuanhua
AU - Yin, Wenpeng
AU - Feng, Xiaocheng
AU - Qin, Bing
AU - Liu, Ting
N1 - Publisher Copyright:
© 2020 IEEE.
PY - 2020/5/1
Y1 - 2020/5/1
N2 - Question answering (QA) and question generation (QG) are closely related tasks that could improve each other; however, the connection of these two tasks is not well explored in the literature. In this paper, we present two training algorithms for learning better QA and QG models through leveraging one another. The first algorithm extends Generative Adversarial Network (GAN), which selectively incorporates artificially generated instances as additional QA training data. The second algorithm is an extension of dual learning, which incorporates the probabilistic correlation of QA and QG as additional regularization in training objectives. To test the scalability of our algorithms, we conduct experiments on both document based and table based question answering tasks. Results show that both algorithms improve a QA model in terms of accuracy and QG model in terms of BLEU score. Moreover, we find that the performance of a QG model could be easily improved by a QA model via policy gradient, however, directly applying GAN that regards all the generated questions as negative instances could not improve the accuracy of the QA model. Our algorithm that selectively assigns labels to generated questions would bring a performance boost.
AB - Question answering (QA) and question generation (QG) are closely related tasks that could improve each other; however, the connection of these two tasks is not well explored in the literature. In this paper, we present two training algorithms for learning better QA and QG models through leveraging one another. The first algorithm extends Generative Adversarial Network (GAN), which selectively incorporates artificially generated instances as additional QA training data. The second algorithm is an extension of dual learning, which incorporates the probabilistic correlation of QA and QG as additional regularization in training objectives. To test the scalability of our algorithms, we conduct experiments on both document based and table based question answering tasks. Results show that both algorithms improve a QA model in terms of accuracy and QG model in terms of BLEU score. Moreover, we find that the performance of a QG model could be easily improved by a QA model via policy gradient, however, directly applying GAN that regards all the generated questions as negative instances could not improve the accuracy of the QA model. Our algorithm that selectively assigns labels to generated questions would bring a performance boost.
UR - http://www.scopus.com/inward/record.url?scp=85082823093&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85082823093&partnerID=8YFLogxK
U2 - 10.1109/TKDE.2019.2897773
DO - 10.1109/TKDE.2019.2897773
M3 - Article
AN - SCOPUS:85082823093
SN - 1041-4347
VL - 32
SP - 971
EP - 982
JO - IEEE Transactions on Knowledge and Data Engineering
JF - IEEE Transactions on Knowledge and Data Engineering
IS - 5
M1 - 8636251
ER -