Section 01
Introduction: MPU Framework—A Privacy-Preserving Knowledge Unlearning Solution for Large Language Models
This article introduces the MPU (Multiple Perturbed Copies Unlearning) framework, an algorithm-agnostic privacy-preserving multiple perturbed copies unlearning framework designed to address the dual non-disclosure constraints in knowledge unlearning for large language models (servers are unwilling to share original model parameters, and clients are unwilling to expose unlearning datasets). Through server-side preprocessing (generating perturbed copies) and postprocessing (aggregation and denoising) modules, MPU achieves efficient knowledge unlearning while protecting model parameters and the privacy of unlearning data.