TY - GEN
T1 - Learning to Select from Multiple Options
AU - Du, Jiangshu
AU - Yin, Wenpeng
AU - Xia, Congying
AU - Yu, Philip S.
N1 - Publisher Copyright:
Copyright © 2023, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.
PY - 2023/6/27
Y1 - 2023/6/27
N2 - Many NLP tasks can be regarded as a selection problem from a set of options, e.g., classification tasks, multi-choice QA, etc. Textual entailment (TE) has been shown as the state-of-the-art (SOTA) approach to dealing with those selection problems. TE treats input texts as premises (P), options as hypotheses (H), then handles the selection problem by modeling (P, H) pairwise. Two limitations: (i) the pairwise modeling is unaware of other options, which is less intuitive since humans often determine the best options by comparing competing candidates; (ii) the inference process of pairwise TE is time-consuming, especially when the option space is large. To deal with the two issues, this work first proposes a contextualized TE model (Context-TE) by appending other k options as the context of the current (P, H) modeling. Context-TE is able to learn more reliable decision for the H since it considers various context. Second, we speed up Context-TE by coming up with Parallel-TE, which learns the decisions of multiple options simultaneously. Parallel-TE significantly improves the inference speed while keeping comparable performance with Context-TE. Our methods are evaluated on three tasks (ultra-fine entity typing, intent detection and multi-choice QA) that are typical selection problems with different sizes of options. Experiments show our models set new SOTA performance; particularly, Parallel-TE is faster than the pairwise TE by k times in inference.
AB - Many NLP tasks can be regarded as a selection problem from a set of options, e.g., classification tasks, multi-choice QA, etc. Textual entailment (TE) has been shown as the state-of-the-art (SOTA) approach to dealing with those selection problems. TE treats input texts as premises (P), options as hypotheses (H), then handles the selection problem by modeling (P, H) pairwise. Two limitations: (i) the pairwise modeling is unaware of other options, which is less intuitive since humans often determine the best options by comparing competing candidates; (ii) the inference process of pairwise TE is time-consuming, especially when the option space is large. To deal with the two issues, this work first proposes a contextualized TE model (Context-TE) by appending other k options as the context of the current (P, H) modeling. Context-TE is able to learn more reliable decision for the H since it considers various context. Second, we speed up Context-TE by coming up with Parallel-TE, which learns the decisions of multiple options simultaneously. Parallel-TE significantly improves the inference speed while keeping comparable performance with Context-TE. Our methods are evaluated on three tasks (ultra-fine entity typing, intent detection and multi-choice QA) that are typical selection problems with different sizes of options. Experiments show our models set new SOTA performance; particularly, Parallel-TE is faster than the pairwise TE by k times in inference.
UR - http://www.scopus.com/inward/record.url?scp=85162756867&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85162756867&partnerID=8YFLogxK
U2 - 10.1609/aaai.v37i11.26500
DO - 10.1609/aaai.v37i11.26500
M3 - Conference contribution
AN - SCOPUS:85162756867
T3 - Proceedings of the 37th AAAI Conference on Artificial Intelligence, AAAI 2023
SP - 12754
EP - 12762
BT - AAAI-23 Technical Tracks 11
A2 - Williams, Brian
A2 - Chen, Yiling
A2 - Neville, Jennifer
PB - AAAI press
T2 - 37th AAAI Conference on Artificial Intelligence, AAAI 2023
Y2 - 7 February 2023 through 14 February 2023
ER -