Section 01
[Main Floor/Introduction] Latent Circuit Disruption: A New Robust Unlearning Method for Large Language Models
This article introduces a model unlearning technique called Latent Circuit Disruption (LCD), whose core is to achieve secure deletion of sensitive information by precisely locating and modifying specific knowledge circuits in large language models while preserving other capabilities of the model. Compared to traditional methods, LCD has significant advantages in unlearning completeness, side effect control, and robustness, providing a new direction for privacy protection and controllability of large language models.