-
-
Notifications
You must be signed in to change notification settings - Fork 7.8k
ImportError: libcudart.so.11.0: cannot open shared object file: No such file or directory #1369
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I encountered the same problem, did you solve it? |
Unfortunately, I didn't find a better solution than installing CUDA 11, but I don't want to make any change about the CUDA version since the machine is not my private, and re-installing CUDA often cause many more unexpected problems. If you have got any solution, please tell me, much appreciated. |
I tried with cuda 12.2. I get the same error, when trying with cuda 11.7 getting the following error : i updated my xformers with pip install xformers==v0.0.22 and works fine. i am using cuda11.7 docker image. |
i have sloved it. first find the libcudart.so.11.0 path on your disk.then write it into LD_LIBRARY_PATH locate libcudart.so.11.0 |
We are getting the same "error", but with CUDA I am not sure who's fault is it, but throwing an error for the reason that "we cannot find a file we installed ourselves so we crash everything" is a bit ridiculous . I did not see any restriction against using CUDA 12 in vLLM docs. So we can expect vLLM works for the latest CUDA version Here is the code to reproduce (cf. below to see in which docker image to run this) from langchain.chains import ConversationChain
from langchain.memory import ConversationBufferMemory
from langchain.llms import VLLM
llm = VLLM(
model="mistralai/Mistral-7B-Instruct-v0.1",
max_new_tokens=8000,
top_k=10,
top_p=0.95,
temperature=0.8,
)
conversation = ConversationChain(
llm=llm, verbose=True, memory=ConversationBufferMemory()
)
print(conversation.predict(input="Hi mom!")) Here is the full error we get:
1. We use official nvidia imagesHere is our setup: We are literally using the official CUDA image from nvidia: Starting from that the message "can't find libcudart" has no reason to exist 2. We made sure to have the right CUDA versionWe don't use the pytorch one like recommended by vllm docs because with the pytorch one we can't control exactly which CUDA version gets installed, and then we get error like "muuuuh pytorch was compiled with a different CUDA version" . Also pytorch GPU image is like ~9G vs ~3G for CUDA
| NVIDIA-SMI 530.30.02 Driver Version: 530.30.02 CUDA Version: 12.1 |
nvidia-cuda-cupti-cu12 12.1.105
nvidia-cuda-nvrtc-cu12 12.1.105
nvidia-cuda-runtime-cu12 12.1.105 3. It was working on CUDA 12.1 last week 🫠We managed to make it work last week when running in a old pytorch docker image that was still on py 3.8. But now it is broken when running on up-to-date images (what a mess), always complaining about this non-existing error with libcudart location And last week when it was working the main GPU was still on CUDA 12.1 (according to
Meaning vLLM can work on CUDA 12 drivers, and we don't need to reinstall CUDA 11, only some CUDA 11 runtime libs should be good: nvidia-cuda-runtime-cu11, or nvidia-cuda-nvrtc-cu11, or nvidia-cuda-cupti-cu11 4. Trying the
|
A potential approach to fix it: it could be due to the Because I'll try to re-build |
let me know if you've got any news here. I've got the same problem since this morning, with nvidia image as well |
It's weird because the latest vllm release actually uses But installing vllm always installs torch 2.0.1, and it due to:
If we try to
But the And whatever combination tried I am always gettings errors, most of the time this one:
|
If I understand the problem correctly, the issue was because v0.2.0 didn't fix the pytorch and xformers version. In v0.2.1 which was released today, we pinned their versions. So the error should not happen as long as you use CUDA 11.8. We will support CUDA 12 once |
Right now Pytorch 2.0.1 is binded to CUDA 11.7, compile vLLM with CUDA 11.8 will get fialed, just the same issue like #1283 |
Finally I installed CUDA 11.7 manually and the problem got fixed immediately. It seemed that vllm cannot work if there is only CUDA 12 installed on the machine. |
I encountered the same problem,
and I solved this error with
|
I'm getting the same error on colab with |
@bitsnaps same problem here. did you find any solution? |
Not yet, I believe this is something to do with mistral/transformer/huggingface issue (not vllm), I'm not even able to run mistral-7b on colab which was working fine last week. |
@bitsnaps I tried to run the Mistral_7B_Instruct_v0_1_GGUF now and its working. I just downgraded gradio to |
Currently, AutoAWQ deliver two versions. (cuda11 and cuda12)
from github(torch20 + cuda11) |
Have a go at updating vllm to v0.2.2. Looks like they've sorted out this issue in that version. |
I'm using llmware library and was facing the same error. I upgraded toch (2.0.1 -> 2.1.0) and solved the prob. |
Thank you! Worked like a charm! |
It's probably just a cuda or torch version problem, try downgrading it |
|
Uh oh!
There was an error while loading. Please reload this page.
When I used vllm to serve my local model, the terminal displayed the following message:
ImportError: libcudart.so.11.0: cannot open shared object file: No such file or directory
The traceback pointed to the following code in site-packages/vllm/utils.py and the execution of the single line could also trigger the same error:
"from vllm import cuda_utils"
I suppose it may be caused by the mismatch between vllm and my CUDA version or Pytorch version. The CUDA version is 12.2 (only this version installed) on my machine and installing a new version 11 is not so convenient, the Pytorch version is 2.1.0, vllm version is 0.2.0
How could I solve the problem without re-install CUDA 11?
Many thanks!
The text was updated successfully, but these errors were encountered: