Replies: 4 comments 3 replies
-
I think the short answer is no, as vLLM's engine relies on custom kernels written in CUDA. |
Beta Was this translation helpful? Give feedback.
-
You can try ctranslate2 or llama.cpp. |
Beta Was this translation helpful? Give feedback.
-
This was not clear to me either - any way to highlight this in bold somewhere on main docs? Sorry if I overlooked. I am trying to do some local testing - that's my use case |
Beta Was this translation helpful? Give feedback.
-
Intel CPU is supported from my own experience. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Can we use vllm only on CPU without GPU machine?
Beta Was this translation helpful? Give feedback.
All reactions