You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
As per title, really. Running a pretty basic ollama pod and then trying to run a model within that results in a segmentation fault on starting the server.
I've tried to hunt for the reason and the strace output seems to contain hints of it running out of memory
Try setting /proc/sys/vm/max_map_count to 1000000. If that fixes things, then this is caused by a known issue. runsc-sandbox exhausts host VMAs while dealing with such large model files in memory.
Description
As per title, really. Running a pretty basic
ollama
pod and then trying to run a model within that results in a segmentation fault on starting the server.I've tried to hunt for the reason and the strace output seems to contain hints of it running out of memory
However, the process should have plenty of memory available, so I seem to be missing something.
Steps to reproduce
I'm testing this in a Kubernetes pod, like so
I then exec into the pod an do
ollama run granite3-dense
(for example, the model doesn't seem to matter, I tried a few) and getThe gvisor setup is as per https://gvisor.dev/docs/user_guide/containerd/configuration/.
runsc version
docker version (if using docker)
No response
uname
Linux pool-2i3vypv5x-g3267 6.1.0-25-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.1.106-3 (2024-08-26) x86_64 GNU/Linux
kubectl (if using Kubernetes)
repo state (if built from source)
No response
runsc debug logs (if available)
debug-strace.log.zip
The text was updated successfully, but these errors were encountered: