Skip to content

xzhou98/Unlearning-TPO

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

29 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Not All Tokens Are Meant to Be Forgotten (AAAI-2026 Oral)

Venue: AAAI 2026 preprint issues GitHub repo size GitHub top language GitHub stars

This is the official code repository for AAAI 2026 paper "Not All Tokens Are Meant to Be Forgotten" by Xiangyu Zhou, Yao Qiang, Saleh Zare Zade, Douglas Zytko, Prashant Khanduri, Dongxiao Zhu

Teaser
Figure 1: Illustration of the proposed TIF framework.

News

📣 Check out our ICLR 2026 paper on improving the robustness of LLM unlearning through attention smoothing!

Abstract

Large Language Models (LLMs), pre-trained on massive text corpora, exhibit remarkable human-level language understanding, reasoning, and decision-making abilities. However, they tend to memorize unwanted information, such as private or copyrighted content, raising significant privacy and legal concerns. Unlearning has emerged as a promising solution, but existing methods face a significant challenge of over-forgetting. This issue arises because they indiscriminately suppress the generation of all the tokens in forget samples, leading to a substantial loss of model utility. To overcome this challenge, we introduce the Targeted Information Forgetting (TIF) framework, which consists of (1) a flexible targeted information identifier designed to differentiate between unwanted words (UW) and general words (GW) in the forget samples, and (2) a novel Targeted Preference Optimization approach that leverages Logit Preference Loss to unlearn unwanted information associated with UW and Preservation Loss to retain general information in GW, effectively improving the unlearning process while mitigating utility degradation. Extensive experiments on the TOFU and MUSE benchmarks demonstrate that the proposed TIF framework enhances unlearning effectiveness while preserving model utility and achieving state-of-the-art results.

Getting Started

Cite This Work

@article{zhou2025not,
  title={Not All Tokens Are Meant to Be Forgotten},
  author={Zhou, Xiangyu and Qiang, Yao and Zade, Saleh Zare and Zytko, Douglas and Khanduri, Prashant and Zhu, Dongxiao},
  journal={arXiv preprint arXiv:2506.03142},
  year={2025}
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors