-
-
Notifications
You must be signed in to change notification settings - Fork 7.4k
[Bug]: RuntimeError: Failed to infer device type
with v0.7.2
#12847
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
looks strange. I think #12809 should solve it. the actual error should raise earlier. |
this is the problem. can you try to follow Line 2223 in b260782
since you have both |
升级pynvml到12.0.0即可。 pip install -U pynvml |
@lizongyao123 @imangohari1 can you install the latest release (0.7.2 has been released) and see if it happens again? I think the latest release should work even if you have |
I have same problem: No platform detected, vLLM is running on UnspecifiedPlatform After ungrade pynvml to 12.0.0 |
It may not work when I use vllm=0.7.2 & pynvml==11.5.3 |
I encountered similar problem |
Very good issue,I also succeeded to fix this by |
@ghosthamlet are you using |
If anyone can give me a reproducible example, that would be great. In addition, feel free to join https://slack.vllm.ai for quick communication. |
I have included a reproducer in the #12847 (comment). |
@youkaichao I redid the test with Note The only workaround right now is to update pynvml to 12.0.0 with |
@imangohari1 thanks! I find the root cause, the docker image |
Thanks, Two points here:
|
same error,really work by update “pip install pynvml==12.0.0” |
It works for me, “pip install pynvml==12.0.0” |
@youkaichao should we pin the |
I found the culprint: when To avoid the issue completely, I opened #12963 . |
@Stonesjtu we cannot pin |
@youkaichao
Should I open a separate ticket for this? |
I opened a new issue for this. #13047 |
Hi, i just download vllm on a ubuntu linux 11feb 11:32 spanish hour, and it just happen the same error. |
Exact same issue with
|
Getting the same error trying to run on Gaudi HPU. |
the same error on V100. |
…0785) Two dependencies we are resolving that requires to pin `xgrammar` and `pynvml` to specific versions. Related vllm PR/ issues - vllm-project/vllm#13338 - vllm-project/vllm#12847 --------- Signed-off-by: Gene Su <[email protected]>
…y-project#50785) Two dependencies we are resolving that requires to pin `xgrammar` and `pynvml` to specific versions. Related vllm PR/ issues - vllm-project/vllm#13338 - vllm-project/vllm#12847 --------- Signed-off-by: Gene Su <[email protected]> Signed-off-by: kaihsun <[email protected]>
I'm seeing the same thing on ROCm after building from source. Will try an earlier version. Update: |
…y-project#50785) Two dependencies we are resolving that requires to pin `xgrammar` and `pynvml` to specific versions. Related vllm PR/ issues - vllm-project/vllm#13338 - vllm-project/vllm#12847 --------- Signed-off-by: Gene Su <[email protected]>
…y-project#50785) Two dependencies we are resolving that requires to pin `xgrammar` and `pynvml` to specific versions. Related vllm PR/ issues - vllm-project/vllm#13338 - vllm-project/vllm#12847 --------- Signed-off-by: Gene Su <[email protected]>
for people who met this issue, I added lots of debug logging in #14195 . please set |
…y-project#50785) Two dependencies we are resolving that requires to pin `xgrammar` and `pynvml` to specific versions. Related vllm PR/ issues - vllm-project/vllm#13338 - vllm-project/vllm#12847 --------- Signed-off-by: Gene Su <[email protected]>
…y-project#50785) Two dependencies we are resolving that requires to pin `xgrammar` and `pynvml` to specific versions. Related vllm PR/ issues - vllm-project/vllm#13338 - vllm-project/vllm#12847 --------- Signed-off-by: Gene Su <[email protected]> Signed-off-by: Jay Chia <[email protected]>
…y-project#50785) Two dependencies we are resolving that requires to pin `xgrammar` and `pynvml` to specific versions. Related vllm PR/ issues - vllm-project/vllm#13338 - vllm-project/vllm#12847 --------- Signed-off-by: Gene Su <[email protected]> Signed-off-by: Jay Chia <[email protected]>
@youkaichao i see the following in the debug log for vllm/vllm-openai:v0.8.1 image.
it looks like i need to build cpu image for vllm/vllm-openai:v0.8.1 myself - is there some doc to build the image? |
no. |
same in h200 with BAAI/bge-m3, but works fine with llms like QwQ:
docker logs:
docker images from |
I think I have found the problem, I have upgraded my docker, and now I need to remove the colon in docker-compose.
Everything is normal now. |
I solved by manually designating the device type in initialization llm = LLM('model_name', device=cuda) |
Your current environment
🐛 Describe the bug
Using the main branch on commit
467a96a5415dc896170cecc0bb83d9c49c2f3c5e
, we are facingRuntimeError: Failed to infer device type
. The details of the env/test is given below.Note
This issue does NOT happen using 0.7.1 release. On the same machine, same container, changing the installation to
pip install vllm
(orpip install https://github.com/vllm-project/vllm/releases/download/v0.7.1/vllm-0.7.1-cp38-abi3-manylinux1_x86_64.whl
) works fine.Container/Setup- Container:
- Setup:
nvcr.io/nvidia/pytorch:24.12-py3
Test
- Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: