Error: llama runner process has terminated: exit status 2 when running ollama
#1
by
DuyDoanLearning
- opened
I run this command:
ollama run hf.co/openbmb/MiniCPM-o-2_6-gguf:Q4_K_M
and has this error:
Error: llama runner process has terminated: exit status 2
Specs:
- CUDA: v12
- GPU T4 Tesla
- Running in SageMaker instance (AWS)
- Python 3.11.11
Model has been downloaded but failed to run. I tried ollama list and confirmed that model has been downloaded:
hf.co/openbmb/MiniCPM-o-2_6-gguf:Q4_K_M 577cda2e3886 5.7 GB 21 seconds ago
I also tried running on llama-cpp-python but it also has a lot of errors.
Please help. I am desperate.
Thank you very much.
I get an error message as well running this on Ollama. I'm not desperate though, I hope this can be solved in the coming weeks. π