Download

NPU Benchmark is currently in development. Release date is to be determined. Follow the project on GitHub for the latest updates.

🚧 In Development

Release Date: TBD

We are building and validating the NPUBench Standard v1.0 methodology across three platforms. When ready, the tool will be released as open source under Apache 2.0.

3 Platforms
Apache 2.0
Windows 11

System Requirements

From NPUBench Standard v1.0

Qualcomm
Snapdragon X Series
OS Windows 11 ARM64
Min Driver ≥ 30.0.140.1000
Min Memory 32 GB
Execution Provider NativeQNN
AMD
Ryzen AI 300 Series (XDNA2+)
OS Windows 11 x64
Min Driver ≥ 32.0.203.280
Min Memory 32 GB
Execution Provider OrtGenAI-RyzenAI
Intel
Core Ultra Series 2
OS Windows 11 x64
Min Driver ≥ 32.0.100.4297
Min Memory 16 GB
Execution Provider NativeOpenVINO

Driver version requirements sourced from MLPerf Client v1.5 reference data.

Benchmark Models

From NPUBench Standard v1.0 Section 6

Model Quantization License
Llama 3.2 3B Instruct int4 Meta Llama 3 Community License
Llama 3.1 8B Instruct int4 Meta Llama 3 Community License
Phi-3.5 Mini Instruct int4 MIT License
⚠ Llama model results must not be used to train or improve non-Llama AI models. See Meta Llama 3 Community License Agreement.

Stay Updated

Follow the project on GitHub for development updates, or join the NPU GenAI developer community to discuss NPU inference and benchmarking.