TY - GEN
T1 - Robust Unlearning for Large Language Models
AU - Gu, Kang
AU - Rashid, Md Rafi Ur
AU - Sultana, Najrin
AU - Mehnaz, Shagufta
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.
PY - 2025
Y1 - 2025
N2 - With the rapid development of LLMs, we have witnessed intense competition among the major LLM products like ChatGPT, LLaMa, and Gemini. However, various issues (e.g. privacy leakage and copyright violation) of the training corpus still remain underexplored. For example, the Times sued OpenAI and Microsoft for infringing on its copyrights by using millions of its articles for training. From the perspective of LLM practitioners, handling such unintended privacy violations can be challenging. Previous work mainly approached the “unlearning" problem of LLMs via gradient information, while they mostly lacked theoretical guarantees. In this paper, we revisit the unlearning via the perspective of second-order information (Hessian). Our unlearning algorithms, inspired by the classic Newton update, are not only data-agnostic/model-agnostic but can also derive an upper bound for utility or privacy loss. Through a comprehensive evaluation with common NLP datasets and case studies on real-world datasets, our methods consistently show superiority over first-order methods.
AB - With the rapid development of LLMs, we have witnessed intense competition among the major LLM products like ChatGPT, LLaMa, and Gemini. However, various issues (e.g. privacy leakage and copyright violation) of the training corpus still remain underexplored. For example, the Times sued OpenAI and Microsoft for infringing on its copyrights by using millions of its articles for training. From the perspective of LLM practitioners, handling such unintended privacy violations can be challenging. Previous work mainly approached the “unlearning" problem of LLMs via gradient information, while they mostly lacked theoretical guarantees. In this paper, we revisit the unlearning via the perspective of second-order information (Hessian). Our unlearning algorithms, inspired by the classic Newton update, are not only data-agnostic/model-agnostic but can also derive an upper bound for utility or privacy loss. Through a comprehensive evaluation with common NLP datasets and case studies on real-world datasets, our methods consistently show superiority over first-order methods.
UR - https://www.scopus.com/pages/publications/105009406982
UR - https://www.scopus.com/inward/citedby.url?scp=105009406982&partnerID=8YFLogxK
U2 - 10.1007/978-981-96-8186-0_12
DO - 10.1007/978-981-96-8186-0_12
M3 - Conference contribution
AN - SCOPUS:105009406982
SN - 9789819681853
T3 - Lecture Notes in Computer Science
SP - 143
EP - 155
BT - Advances in Knowledge Discovery and Data Mining - 29th Pacific-Asia Conference on Knowledge Discovery and Data Mining, PAKDD 2025, Proceedings
A2 - Wu, Xintao
A2 - Spiliopoulou, Myra
A2 - Wang, Can
A2 - Kumar, Vipin
A2 - Cao, Longbing
A2 - Wu, Yanqiu
A2 - Wu, Zhangkai
A2 - Yao, Yu
PB - Springer Science and Business Media Deutschland GmbH
T2 - 29th Pacific-Asia Conference on Knowledge Discovery and Data Mining, PAKDD 2025
Y2 - 10 June 2025 through 13 June 2025
ER -