Section 01
Watt Counts: A Guide to LLM Energy Efficiency Optimization Under Heterogeneous GPU Architectures (Introduction)
Watt Counts is a guide project focused on energy efficiency optimization for Large Language Models (LLMs) under heterogeneous GPU architectures. It provides over 5000 experimental data points (covering 50 LLMs and 10 NVIDIA GPU models) and reveals the critical impact of hardware selection on energy efficiency. This project helps practitioners reduce energy consumption by 70% in server-side scenarios and 20% in batch processing scenarios, filling the gap in system-level energy-aware benchmarking and datasets.