-
Notifications
You must be signed in to change notification settings - Fork 5.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ollama run codellama:34b issue #4519
Comments
same here |
So sorry about this, looking into why this is happening. |
same |
I'm also getting this on 7b-python and 13b-python |
Could you share your server log, and your VRAM size? This is a larger model, so my suspicion is we're off on our memory prediction. |
Same here. AMD CPU in contrast to OP. 16 GB VRAM (RTX 4070 Ti Super) Log attached. |
To add details .. AMD 5950X on x570E 6900XT server log; doesn't appear to have the events in it anymore, but I'll try and recreate later today if I can. |
Hereby my log. I can run mixtral and dolphin-mixtral fine and those are some big ones too. Anyhow my current system is i7-9700K + 32 GB RAM + RTX 3090 with 24 GB VRAM |
I7 7700K with 32gb ram and 6 GPUs with in total 51gb vram. I can run all models, even llama3:70b without problems, which is already a big model. But I get error with codellama 34b |
same here .I use 0.1.38 version of official docker image.
Gentleman, I find another way, downgrade to 0.1.32 is okay to load codellama and large model. 0.1.33 - 0.1.39 I tested all not worked. |
same. Phind-codellama from the repository works, loading gguf doesn't though |
What is the issue?
Every model I tested with ollama runs fine, but when trying:
ollama run codellama:34b
, I getError: llama runner process has terminated: signal: segmentation fault (core dumped)
Tried the 13B-version then, works fine.OS
Linux
GPU
Nvidia
CPU
Intel
Ollama version
0.1.37
The text was updated successfully, but these errors were encountered: