
Sign up to save your podcasts
Or
Seventy3: 用NotebookLM将论文生成播客,让大家跟着AI一起进步。
今天的主题是:Does your LLM truly unlearn? An embarrassingly simple approach to recover unlearned knowledgeSummary
This research paper investigates a critical flaw in current machine unlearning methods for large language models (LLMs). The authors discover that applying quantization, a process used to compress and optimize LLMs for resource-constrained environments, can inadvertently restore "forgotten" knowledge. The paper provides a theoretical explanation for this phenomenon and proposes a new unlearning strategy, "Saliency-Based Unlearning with a Large Learning Rate (SURE)," to mitigate this issue and ensure genuine unlearning without compromising model utility. The study underscores the need for more comprehensive and robust approaches to machine unlearning in LLMs, highlighting a critical oversight in existing unlearning benchmarks.
原文链接:https://arxiv.org/abs/2410.16454
解读链接:https://www.qbitai.com/2024/11/219654.html
Seventy3: 用NotebookLM将论文生成播客,让大家跟着AI一起进步。
今天的主题是:Does your LLM truly unlearn? An embarrassingly simple approach to recover unlearned knowledgeSummary
This research paper investigates a critical flaw in current machine unlearning methods for large language models (LLMs). The authors discover that applying quantization, a process used to compress and optimize LLMs for resource-constrained environments, can inadvertently restore "forgotten" knowledge. The paper provides a theoretical explanation for this phenomenon and proposes a new unlearning strategy, "Saliency-Based Unlearning with a Large Learning Rate (SURE)," to mitigate this issue and ensure genuine unlearning without compromising model utility. The study underscores the need for more comprehensive and robust approaches to machine unlearning in LLMs, highlighting a critical oversight in existing unlearning benchmarks.
原文链接:https://arxiv.org/abs/2410.16454
解读链接:https://www.qbitai.com/2024/11/219654.html