The content discusses the importance of optimizer choice in large language model (LLM) unlearning, which aims to remove undesired data influences and associated model capabilities without compromising utility.
The key highlights are:
The authors establish a clear connection between second-order optimization and influence unlearning, a classical approach that uses influence functions to update the model for data influence removal.
Motivated by this insight, the authors propose SOUL, a second-order unlearning framework built upon the second-order clipped stochastic optimization (Sophia) method. SOUL extends the static, one-shot model update using influence unlearning to a dynamic, iterative unlearning process.
Extensive experiments across various unlearning tasks, models, and metrics consistently show that SOUL outperforms conventional first-order methods, suggesting the promise of second-order optimization in providing a scalable and easily implementable solution for LLM unlearning.
The authors demonstrate that SOUL can effectively remove undesired data influences, such as fictitious author information, copyrighted content, and toxic language, while preserving the model's utility for unrelated tasks.
The results advocate for the development and adoption of optimizers tailored for effective LLM unlearning, as the choice of optimizer plays a crucial role in the unlearning process.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Jinghan Jia,... at arxiv.org 04-30-2024
https://arxiv.org/pdf/2404.18239.pdfDeeper Inquiries