-
Notifications
You must be signed in to change notification settings - Fork 280
NPU Not Detected foundry model list Only Shows CPU Models #584
Copy link
Copy link
Open
Description
I’m on Windows using the recommended installer. I have an NPU‑equipped device, but when I run:
foundry model list
every model is listed as CPU‑only. Nothing shows up as NPU/DML‑accelerated.
PS C:\Users\luongdavid> foundry model list
Alias Device Task File Size License Model ID
-----------------------------------------------------------------------------------------------
phi-4 CPU chat 10.16 GB MIT Phi-4-generic-cpu:1
----------------------------------------------------------------------------------------------------------
phi-3.5-mini CPU chat 2.53 GB MIT Phi-3.5-mini-instruct-generic-cpu:1
--------------------------------------------------------------------------------------------------------------------------
phi-3-mini-128k CPU chat 2.54 GB MIT Phi-3-mini-128k-instruct-generic-cpu:2
-----------------------------------------------------------------------------------------------------------------------------
phi-3-mini-4k CPU chat 2.53 GB MIT Phi-3-mini-4k-instruct-generic-cpu:2
---------------------------------------------------------------------------------------------------------------------------
mistral-7b-v0.2 CPU chat 4.07 GB apache-2.0 mistralai-Mistral-7B-Instruct-v0-2-generic-cpu:2
---------------------------------------------------------------------------------------------------------------------------------------
deepseek-r1-14b CPU chat 11.51 GB MIT deepseek-r1-distill-qwen-14b-generic-cpu:3
---------------------------------------------------------------------------------------------------------------------------------
deepseek-r1-7b CPU chat 6.43 GB MIT deepseek-r1-distill-qwen-7b-generic-cpu:3
--------------------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-0.5b CPU chat, tools 0.80 GB apache-2.0 qwen2.5-coder-0.5b-instruct-generic-cpu:4
--------------------------------------------------------------------------------------------------------------------------------
phi-4-mini-reasoning CPU chat 4.52 GB MIT Phi-4-mini-reasoning-generic-cpu:3
-------------------------------------------------------------------------------------------------------------------------
qwen2.5-0.5b CPU chat, tools 0.80 GB apache-2.0 qwen2.5-0.5b-instruct-generic-cpu:4
--------------------------------------------------------------------------------------------------------------------------
qwen2.5-1.5b CPU chat, tools 1.78 GB apache-2.0 qwen2.5-1.5b-instruct-generic-cpu:4
--------------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-1.5b CPU chat, tools 1.78 GB apache-2.0 qwen2.5-coder-1.5b-instruct-generic-cpu:4
--------------------------------------------------------------------------------------------------------------------------------
phi-4-mini CPU chat, tools 4.80 GB MIT Phi-4-mini-instruct-generic-cpu:5
------------------------------------------------------------------------------------------------------------------------
qwen2.5-14b CPU chat, tools 11.06 GB apache-2.0 qwen2.5-14b-instruct-generic-cpu:4
-------------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-14b CPU chat, tools 11.06 GB apache-2.0 qwen2.5-coder-14b-instruct-generic-cpu:4
-------------------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-7b CPU chat, tools 6.16 GB apache-2.0 qwen2.5-coder-7b-instruct-generic-cpu:4
------------------------------------------------------------------------------------------------------------------------------
qwen2.5-7b CPU chat, tools 6.16 GB apache-2.0 qwen2.5-7b-instruct-generic-cpu:4
------------------------------------------------------------------------------------------------------------------------
gpt-oss-20b CPU chat 12.26 GB MIT gpt-oss-20b-generic-cpu:1
----------------------------------------------------------------------------------------------------------------
qwen3-0.6b CPU chat, tools 0.58 GB apache-2.0 qwen3-0.6b-generic-cpu:3
Environment
I am on a 24H2, following these steps : #510 (comment)
Foundry Local version: 0.8.119
Hardware:
NPU:
Snapdragon(R) X Elite - X1E78100 - Qualcomm(R) Hexagon(TM) NPU
Driver version: 30.0.220.3000
Driver date: 1/8/2026
I did run Get-AppxPackage -AllUsers "*.EP.*" | Select-Object -ExpandProperty PackageFullName
I do have 4 qnn providers present.
Note : I am internal msft and self hosting internal branches.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels