상세 보기
- Kwon, Junehyoung;
- Kim, Mihyeon;
- Lee, Eunju;
- Lee, Yoonji;
- Lee, Seunghoon;
- ... Kim, Youngbin
WEB OF SCIENCE
0SCOPUS
0초록
Machine unlearning, which enables a model to forget specific data, is crucial for ensuring data privacy and model reliability. However, its effectiveness can be severely undermined in real-world scenarios where models learn unintended biases from spurious correlations within the data. This paper investigates the unique challenges of unlearning from such biased models. We identify a novel phenomenon we term "shortcut unlearning," where models exhibit an "easy to learn, yet hard to forget" tendency. Specifically, models struggle to forget easily-learned, bias-aligned samples; instead of forgetting the class attribute, they unlearn the bias attribute, which can paradoxically improve accuracy on the class intended to be forgotten. To address this, we propose CUPID, a new unlearning framework inspired by the observation that samples with different biases exhibit distinct loss landscape sharpness. Our method first partitions the forget set into causal- and bias-approximated subsets based on sample sharpness, then disentangles model parameters into causal and bias pathways, and finally performs a targeted update by routing refined causal and bias gradients to their respective pathways. Extensive experiments on biased datasets including Waterbirds, BAR, and Biased NICO++ demonstrate that our method achieves state-of-the-art forgetting performance and effectively mitigates the shortcut unlearning problem.
- 제목
- Easy to Learn, Yet Hard to Forget: Towards Robust Unlearning Under Bias
- 저자
- Kwon, Junehyoung; Kim, Mihyeon; Lee, Eunju; Lee, Yoonji; Lee, Seunghoon; Kim, Youngbin
- 발행일
- 2026
- 유형
- Conference Paper
- 저널명
- Proceedings of the AAAI Conference on Artificial Intelligence
- 권
- 40
- 호
- 7
- 페이지
- 5782 ~ 5790