Section 01
Attention Atlas: An Open-Source Tool for LLM Interpretability Through Attention Visualization
Attention Atlas is an open-source master thesis project that provides a complete toolset for visualizing the attention mechanisms of large language models (LLMs). Its core purpose is to help researchers and developers explore attention patterns, assess model biases, and verify interpretability—addressing the "black box" nature of attention weights, which is crucial for debugging models, identifying biases, and improving performance.