TY - GEN
T1 - Unified Low-Resource Sequence Labeling by Sample-Aware Dynamic Sparse Finetuning
AU - Das, Sarkar Snigdha Sarathi
AU - Zhang, Ranran Haoran
AU - Shi, Peng
AU - Yin, Wenpeng
AU - Zhang, Rui
N1 - Publisher Copyright:
© 2023 Association for Computational Linguistics.
PY - 2023
Y1 - 2023
N2 - Unified Sequence Labeling articulates different sequence labeling tasks such as Named Entity Recognition, Relation Extraction, Semantic Role Labeling, etc. in a generalized sequence-to-sequence format. Unfortunately, this requires formatting different tasks into specialized augmented formats which are unfamiliar to the base pretrained language model (PLMs). This necessitates model fine-tuning and significantly bounds its usefulness in data-limited settings where fine-tuning large models cannot properly generalize to the target format. To address this challenge and leverage PLM knowledge effectively, we propose FISH-DIP, a sample-aware dynamic sparse finetuning strategy. It selectively finetunes a fraction of parameters informed by highly regressing examples during the fine-tuning process. By leveraging the dynamism of sparsity, our approach mitigates the impact of well-learned samples and prioritizes underperforming instances for improvement in generalization. Across five tasks of sequence labeling, we demonstrate that FISH-DIP can smoothly optimize the model in low-resource settings, offering up to 40% performance improvements over full fine-tuning depending on target evaluation settings. Also, compared to in-context learning and other parameter-efficient fine-tuning (PEFT) approaches, FISH-DIP performs comparably or better, notably in extreme low-resource settings. The source code of FISH-DIP will be available at: https://github.com/psunlpgroup/FISH-DIP.
AB - Unified Sequence Labeling articulates different sequence labeling tasks such as Named Entity Recognition, Relation Extraction, Semantic Role Labeling, etc. in a generalized sequence-to-sequence format. Unfortunately, this requires formatting different tasks into specialized augmented formats which are unfamiliar to the base pretrained language model (PLMs). This necessitates model fine-tuning and significantly bounds its usefulness in data-limited settings where fine-tuning large models cannot properly generalize to the target format. To address this challenge and leverage PLM knowledge effectively, we propose FISH-DIP, a sample-aware dynamic sparse finetuning strategy. It selectively finetunes a fraction of parameters informed by highly regressing examples during the fine-tuning process. By leveraging the dynamism of sparsity, our approach mitigates the impact of well-learned samples and prioritizes underperforming instances for improvement in generalization. Across five tasks of sequence labeling, we demonstrate that FISH-DIP can smoothly optimize the model in low-resource settings, offering up to 40% performance improvements over full fine-tuning depending on target evaluation settings. Also, compared to in-context learning and other parameter-efficient fine-tuning (PEFT) approaches, FISH-DIP performs comparably or better, notably in extreme low-resource settings. The source code of FISH-DIP will be available at: https://github.com/psunlpgroup/FISH-DIP.
UR - http://www.scopus.com/inward/record.url?scp=85184828173&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85184828173&partnerID=8YFLogxK
U2 - 10.18653/v1/2023.emnlp-main.433
DO - 10.18653/v1/2023.emnlp-main.433
M3 - Conference contribution
AN - SCOPUS:85184828173
T3 - EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings
SP - 6998
EP - 7010
BT - EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings
A2 - Bouamor, Houda
A2 - Pino, Juan
A2 - Bali, Kalika
PB - Association for Computational Linguistics (ACL)
T2 - 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023
Y2 - 6 December 2023 through 10 December 2023
ER -