Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

NPU Acceleration #505

Open
sikhness opened this issue May 31, 2024 · 4 comments
Open

NPU Acceleration #505

sikhness opened this issue May 31, 2024 · 4 comments
Labels
🔖 ADO Has corresponding ADO item AI/ML Artificial intelligence & machine learning development enhancement New feature or request perf Speed, efficiency, optimization concerns

Comments

@sikhness
Copy link

Currently, only a subset of devices can be passed through, with GPUs being one of them (albeit limited with only DirectX based frameworks).
With the rise and push of NPU/IPUs built into processors, it would be beneficial to provide NPU acceleration in Windows Containers to be able to containerize our AI/ML workloads.

@sikhness sikhness added enhancement New feature or request triage New and needs attention labels May 31, 2024
@ntrappe-msft ntrappe-msft added AI/ML Artificial intelligence & machine learning development perf Speed, efficiency, optimization concerns labels Jun 3, 2024
@fady-azmy-msft
Copy link
Contributor

Hey @sikhness, similar question to your other issue. Can you help me understand what sort of workloads are you trying to run with NPU acceleration? Understanding this use case will help us better prioritize this request as we explore AI/ML workloads.

@sikhness
Copy link
Author

sikhness commented Jun 5, 2024

Hey @fady-azmy-msft!
Similar to my other question, I did list out a few AI related workloads that would benefit from GPU Acceleration from vendor specific graphics APIs.

Some of those same AI workloads can also benefit from offloading that work to the NPU now and here is an example of Ryzen AI which provides instructions on how to install, prep and run your AI models on the NPU on Windows. It would be very beneficial to be able to containerize these applications for isolation & portability benefits and still leverage the hardware.

@fady-azmy-msft fady-azmy-msft removed the triage New and needs attention label Jun 19, 2024
@fady-azmy-msft
Copy link
Contributor

Got it. Tagging @NAWhitehead to look into this. He's driving the Windows containers GPU scenarios, and this is related.

@doctorpangloss
Copy link

I think you should get the class GUID for "Neural processors", try passing it as a --device class/the_guid, copy the drivers from the FileRepository into the container, and then see if the NPU works. Odds are low but crazier things have been true.

@ntrappe-msft ntrappe-msft added the 🔖 ADO Has corresponding ADO item label Sep 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
🔖 ADO Has corresponding ADO item AI/ML Artificial intelligence & machine learning development enhancement New feature or request perf Speed, efficiency, optimization concerns
Projects
None yet
Development

No branches or pull requests

5 participants