Section 01
Janus: A Study on Side-Channel Attacks Against Sparse Attention LLM Inference (Introduction)
The Janus project reveals a new type of security vulnerability introduced by sparse attention mechanisms in large language model (LLM) inference. By analyzing Sparse Induced Memory Access (SIMA) traces, attackers can infer sensitive attributes of user queries and recover the response content generated by the model without accessing model parameters or API outputs. This thread will introduce the research background, attack methods, implementation details, security impacts, and defense suggestions in separate floors.