Back to Explorer
Research PaperResearchia:202602.19006[Data Science > Machine Learning]

Protecting the Undeleted in Machine Unlearning

Aloni Cohen

Abstract

Machine unlearning aims to remove specific data points from a trained model, often striving to emulate "perfect retraining", i.e., producing the model that would have been obtained had the deleted data never been included. We demonstrate that this approach, and security definitions that enable it, carry significant privacy risks for the remaining (undeleted) data points. We present a reconstruction attack showing that for certain tasks, which can be computed securely without deletions, a mechanism adhering to perfect retraining allows an adversary controlling merely ω(1)ω(1) data points to reconstruct almost the entire dataset merely by issuing deletion requests. We survey existing definitions for machine unlearning, showing they are either susceptible to such attacks or too restrictive to support basic functionalities like exact summation. To address this problem, we propose a new security definition that specifically safeguards undeleted data against leakage caused by the deletion of other points. We show that our definition permits several essential functionalities, such as bulletin boards, summations, and statistical learning.


Source: arXiv:2602.16697v1 - http://arxiv.org/abs/2602.16697v1 PDF: https://arxiv.org/pdf/2602.16697v1 Original Link: http://arxiv.org/abs/2602.16697v1

Submission:2/19/2026
Comments:0 comments
Subjects:Machine Learning; Data Science
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Protecting the Undeleted in Machine Unlearning | Researchia | Researchia