Section 01
【Introduction】FHE-native Mamba-3: FHE-native Architecture Unleashes a New Era of Privacy-Preserving LLM Inference
This article introduces the FHE-native Mamba-3 project, which deeply integrates Fully Homomorphic Encryption (FHE) with Mamba's state space model to build a native architecture optimized for encrypted inference. It addresses the inefficiency of traditional Transformers in FHE environments, achieves dual breakthroughs in data privacy and model performance, and paves a new path for privacy-preserving large language model (LLM) inference.