TY - GEN
T1 - All Labels Together
T2 - 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics, IJCNLP-AACL 2023
AU - Du, Jiangshu
AU - Xia, Congying
AU - Yin, Wenpeng
AU - Liang, Tingting
AU - Yu, Philip
N1 - Publisher Copyright:
©2023 Association for Computational Linguistics.
PY - 2023
Y1 - 2023
N2 - In intent detection tasks, leveraging meaningful semantic information from intent labels can be particularly beneficial for few-shot scenarios. However, existing few-shot intent detection methods either ignore the intent labels, (e.g. treating intents as indices) or do not fully utilize this information (e.g. only using part of the intent labels). In this work, we present an end-to-end One-to-All system that enables the comparison of an input utterance with all label candidates. The system can then fully utilize label semantics in this way. Experiments on three few-shot intent detection tasks demonstrate that One-to-All is especially effective when the training resource is extremely scarce, achieving state-of-the-art performance in 1-, 3- and 5-shot settings. Moreover, we present a novel pretraining strategy for our model that utilizes indirect supervision from paraphrasing, enabling zero-shot cross-domain generalization on intent detection tasks. Our code is at https://github.com/jiangshdd/AllLablesTogether.
AB - In intent detection tasks, leveraging meaningful semantic information from intent labels can be particularly beneficial for few-shot scenarios. However, existing few-shot intent detection methods either ignore the intent labels, (e.g. treating intents as indices) or do not fully utilize this information (e.g. only using part of the intent labels). In this work, we present an end-to-end One-to-All system that enables the comparison of an input utterance with all label candidates. The system can then fully utilize label semantics in this way. Experiments on three few-shot intent detection tasks demonstrate that One-to-All is especially effective when the training resource is extremely scarce, achieving state-of-the-art performance in 1-, 3- and 5-shot settings. Moreover, we present a novel pretraining strategy for our model that utilizes indirect supervision from paraphrasing, enabling zero-shot cross-domain generalization on intent detection tasks. Our code is at https://github.com/jiangshdd/AllLablesTogether.
UR - https://www.scopus.com/pages/publications/105027132437
UR - https://www.scopus.com/pages/publications/105027132437#tab=citedBy
U2 - 10.18653/v1/2023.ijcnlp-short.15
DO - 10.18653/v1/2023.ijcnlp-short.15
M3 - Conference contribution
AN - SCOPUS:105027132437
T3 - Proceedings of the 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics: Long Papers, IJCNLP-AACL 2023
SP - 131
EP - 138
BT - Short Papers
A2 - Park, Jong C.
A2 - Arase, Yuki
A2 - Hu, Baotian
A2 - Lu, Wei
A2 - Wijaya, Derry
A2 - Purwarianti, Ayu
A2 - Krisnadhi, Adila Alfa
PB - Association for Computational Linguistics (ACL)
Y2 - 1 November 2023 through 4 November 2023
ER -