Section 01
Introduction / Main Floor: Running Large Language Models on Snapdragon X Elite: Practice of NPU-Accelerated On-Device AI Inference
This article introduces how to run large language model inference on Windows ARM64 devices equipped with Snapdragon X Elite/X2 Elite, using Qualcomm NPU and ONNX Runtime QNN Execution Provider to achieve efficient on-device AI computing.