Few-shot Unlearning
- Title
- Few-shot Unlearning
- Authors
- Youngsik Yoon; Jinhwan Nam; Hyojeong Yun; LEE, JAEHO; KIM, DONGWOO; OK, JUNGSEUL
- Date Issued
- 2024-05-21
- Publisher
- IEEE
- Abstract
- We consider the problem of machine unlearning to erase the impact of a target dataset, used in training but incorrect or sensitive, from a trained model. It has been often presumed that every data sample to erase or remain is entirely identifiable and thus clarifies the desired model behavior after unlearning. However, such a flawless identification can be infeasible in practice. We pose a further realistic yet challenging scenario, referred to as few-shot unlearning, where only a few samples of target data are provided while aiming at achieving the underlying intention (e.g., correcting mislabels, countering a certain privacy attack, or specifying nothing) behind the full target dataset. We then devise a few-shot unlearning method including a new model inversion technique, specialized for unlearning scenarios, to retrieve a proxy of the training dataset from the trained model if needed. We demonstrate that our method using only a tiny subset of target data can achieve similar performance to the state-of-the-art methods with full access to target data.
- URI
- https://oasis.postech.ac.kr/handle/2014.oak/122896
- Article Type
- Conference
- Citation
- IEEE Symposium on Security and Privacy, 2024, 2024-05-21
- Files in This Item:
- There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.