TY - GEN
T1 - NKT
T2 - Joint 30th International Conference on Computational Linguistics and 14th International Conference on Language Resources and Evaluation, LREC-COLING 2024
AU - Feng, Xiaotong
AU - Chiang, Meng Fen
AU - Lee, Wang Chien
AU - Kuang, Zixin
N1 - Publisher Copyright:
© 2024 ELRA Language Resource Association: CC BY-NC 4.0.
PY - 2024
Y1 - 2024
N2 - Cross-domain knowledge transfer, which has received growing research attention in natural language processing (NLP), is a promising approach for various NLP tasks such as evidence-aware inference. However, the presence of biased language in well-known benchmarks notably misleads predictive models due to the hidden false correlations in the linguistic corpus. In this paper, we propose Neutralized Knowledge Transfer framework (NKT) to equip pre-trained language models with neutralized transferability. Specifically, we construct debiased multi-source corpora (CV and EL) for two exemplary knowledge transfer tasks: claim verification and evidence learning, respectively. To counteract biased language, we design a neutralization mechanism in the presence of label skewness. We also design a label adaptation mechanism in light of the mixed label systems in the multi-source corpora. In extensive experiments, the proposed NKT framework shows effective transferability contrarily to the disability of dominant baselines, particularly in the zero-shot cross-domain transfer setting.
AB - Cross-domain knowledge transfer, which has received growing research attention in natural language processing (NLP), is a promising approach for various NLP tasks such as evidence-aware inference. However, the presence of biased language in well-known benchmarks notably misleads predictive models due to the hidden false correlations in the linguistic corpus. In this paper, we propose Neutralized Knowledge Transfer framework (NKT) to equip pre-trained language models with neutralized transferability. Specifically, we construct debiased multi-source corpora (CV and EL) for two exemplary knowledge transfer tasks: claim verification and evidence learning, respectively. To counteract biased language, we design a neutralization mechanism in the presence of label skewness. We also design a label adaptation mechanism in light of the mixed label systems in the multi-source corpora. In extensive experiments, the proposed NKT framework shows effective transferability contrarily to the disability of dominant baselines, particularly in the zero-shot cross-domain transfer setting.
UR - http://www.scopus.com/inward/record.url?scp=85195962000&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85195962000&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85195962000
T3 - 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation, LREC-COLING 2024 - Main Conference Proceedings
SP - 6671
EP - 6681
BT - 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation, LREC-COLING 2024 - Main Conference Proceedings
A2 - Calzolari, Nicoletta
A2 - Kan, Min-Yen
A2 - Hoste, Veronique
A2 - Lenci, Alessandro
A2 - Sakti, Sakriani
A2 - Xue, Nianwen
PB - European Language Resources Association (ELRA)
Y2 - 20 May 2024 through 25 May 2024
ER -