-
-
Notifications
You must be signed in to change notification settings - Fork 7.8k
[Usage]: RuntimeError: Failed to infer device type (Intel Iris Xe Graphics) #8863
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I don't believe we support Intel(R) Iris(R) Xe Graphics with any backend, possibly OpenVINO? By default vLLM is built for NVIDIA GPUs |
A computer restart is needed after vllm installation in most cases. Otherwise the "Failed to infer device type" might display. |
I'm also facing the same issue trying to use the XPU target for Intel Arc A770:
|
I tried OpenVINO but the instructions don't seem to build Intel GPU support, so even in Docker I get all sorts of errors. |
If using CPU, you can specify the device when creating LLM handler. |
Hi @pepijndevos , |
I believe the problem may have been with the version of userland intel drivers. I can run openvino on GPU now. |
This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you! |
This issue has been automatically closed due to inactivity. Please feel free to reopen if you feel it is still relevant. Thank you! |
Uh oh!
There was an error while loading. Please reload this page.
Your current environment
My GPU : Intel(R) Iris(R) Xe Graphics / 15.8GB of shared memory
How would you like to use vllm
I want to run inference of a "mistralai/Pixtral-12B-2409" as shown here. Here below the code :
This is what i got :
Thanks,
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: