-
Notifications
You must be signed in to change notification settings - Fork 64
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
NPU Acceleration #505
Comments
Hey @sikhness, similar question to your other issue. Can you help me understand what sort of workloads are you trying to run with NPU acceleration? Understanding this use case will help us better prioritize this request as we explore AI/ML workloads. |
Hey @fady-azmy-msft! Some of those same AI workloads can also benefit from offloading that work to the NPU now and here is an example of Ryzen AI which provides instructions on how to install, prep and run your AI models on the NPU on Windows. It would be very beneficial to be able to containerize these applications for isolation & portability benefits and still leverage the hardware. |
Got it. Tagging @NAWhitehead to look into this. He's driving the Windows containers GPU scenarios, and this is related. |
I think you should get the class GUID for "Neural processors", try passing it as a |
Currently, only a subset of devices can be passed through, with GPUs being one of them (albeit limited with only DirectX based frameworks).
With the rise and push of NPU/IPUs built into processors, it would be beneficial to provide NPU acceleration in Windows Containers to be able to containerize our AI/ML workloads.
The text was updated successfully, but these errors were encountered: