-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error of running llama3.2-vision with the ipex-llm's built-in ollama. #12707
Comments
The following is the complete and detailed error message:
|
Hi @1ngram433, what's your ollama version? And could you pls provide the specific information of your device? |
Thank you, my ollama version is 0.5.1-ipexllm-20250114, and ipex-llm version is 2.2.0b20250114 AMD Ryzen 5 5500 |
Well, I cannot reproduce your issue. I suspect it is related to the use of an AMD CPU on your device. |
I followed the steps in #https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/ollama_quickstart.md to set up ollama. I can successfully run llama3.2 and qwen. However, when I use
ollama run llama3.2-vision
, I get an error:GGML_ASSERT(ggml_nelements(a) == ne0ne1ne2) failed
time=2025-01-15T10:43:34.636+08:00 level=INFO source=server.go:614 msg="waiting for server to become available" status="llm server error"
time=2025-01-15T10:43:34.887+08:00 level=ERROR source=sched.go:455 msg="error loading llama server" error="llama runner process has terminated: GGML_ASSERT(ggml_nelements(a) == ne0ne1ne2) failed"
The text was updated successfully, but these errors were encountered: