-
-
Notifications
You must be signed in to change notification settings - Fork 7.5k
RuntimeError: attn_bias is not correctly aligned #407
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Labels
bug
Something isn't working
Comments
That is very weird, given I also meet this error, when playing with MPT-7B in my A5000 GPUs. TOT |
76 tasks
metoo... when playing with Baichuan13B on V100 32GB |
groenenboomj
pushed a commit
to opendatahub-io/vllm
that referenced
this issue
Feb 27, 2025
* Update README.md 20250205_aiter * whitespace * adding VLLM_USE_AITER=0 advice
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Unable to handle request for model mosaicml/mpt-30b-chat
Here is my Xformers
Pytorch Version:
The text was updated successfully, but these errors were encountered: