Section 01
Fairness Pruning: A New Method for Mitigating LLM Biases via Activation-Guided MLP Pruning
This article introduces an innovative method called Fairness Pruning, which precisely identifies and removes biased neurons in models through activation-guided MLP width pruning technology. It effectively reduces biases in large language models (LLMs) without significantly sacrificing model performance, providing a new approach to solving the fairness-performance trade-off problem in LLMs.