-
Notifications
You must be signed in to change notification settings - Fork 1.4k
Error: flag provided but not defined: -ngl #13115
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Hi @aarononeal, which version of ipex-llm ollama are you running? |
I see the same when I build the latest xpu-ccp container according to the guide. |
@sgwhat I followed the Docker quickstart using image |
Can also confirm that. |
Confirmed
Gemma3:4b (Q4_0 and QA along with Gemma3:1b) result in error about "-ngl provided but not defined." Also tested models Mistral:7b, llama3.1:8b. Both load and inference on the GPU fine without error. Environment: Docker container built and run exactly as instructed in ipex-llm/docker/llm/inference-cpp. I exported ONEAPI_DEVICE_SELECTOR=level_zero:0 (my GPU) along with OLLAMA_HOST=0.0.0.0 prior to running /llm/scripts/start-ollama.sh. executing /llm/ollama/ollama downloads, loads, and inferences mistral:7b perfectly fine and is visible to open-webui. All attempts at running Gemma3 fail with same error as OP. I even downloaded the raw .gguf from HF, made my own Modelfile and created the model that way, same error when running the model. |
Happens for me too. I just installed ollama-ipex-llm-2.3.0b20250428-win.zip C:>ollama run gemma3 The log shows: But gemma2 works without an issue. |
Same for me, but now for all models.
the difference is Version: |
Ah. Maybe just not supported yet? #12963 |
Hi guys, this version and |
@sgwhat, thank you for your efforts here!
Ah, nevermind, you said tomorrow! I will check back then. Thanks again! |
So I'm happy to report that the These repeat in the logs for a while:
Before a final timeout:
Other entries related to runner and model load:
|
A system restart resolved the above timeouts and the model appears to load. However, now stopped at this issue.
|
@sgwhat: do you know how often the windows nightly builds of ollama are being build? |
For the latest nightly, please refer to https://github.com/intel/ipex-llm/blob/main/docs/mddocs/Quickstart/ollama_quickstart.md |
Although I still can’t get gemma3 working due to the token problem, the specific issue I reported is addressed, so closing this. |
Uh oh!
There was an error while loading. Please reload this page.
Describe the bug
Ollama runner fails with exit code 2 because -ngl is not a parameter.
How to reproduce
Steps to reproduce the error:
Screenshots
Environment information
If possible, please attach the output of the environment check script, using:
Additional context
Note that runner fails because
-ngl
is passed when the parameter allowed is-n-gpu-layers
.The text was updated successfully, but these errors were encountered: