-
-
Notifications
You must be signed in to change notification settings - Fork 1.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Example Chat-UI (ChatGPT OSS Alternative) causing crash of API with preloaded model #574
Comments
Just leaving it here in case others have similar problems ... obvisously my docker-machine had not enough RAM-Memory assigned, causing the crash when trying to load the models into RAM memory. Trying with more memory assigned to the VM and reporting here if it works then. |
Tried with 16 GB RAM attached still crashes the docker-container for localai-api without useful exception pointing out what's going wrong. |
I've cross checked now and deployed the same docker-compose setup on my notebook-workstation (Intel(R) Core(TM) i7-9750H CPU @ 2.60GHz") with Ubuntu OS/Docker. There it works! The previous deployment that caused problems was on my IBM Server which runs VMware ESXi using a Intel(R) Xeon(R) CPU E5620 @ 2.40GHz and Ubuntu OS/Docker VM. So either local-ai stack has any kind of problems with VMware virtualisation or with Intel Xeon CPU or Xeon Model E5620?! |
i have a Xeon E5649 CPU and have the same issue with the api crashing. I suspect it is an incompatible CPU.
my docker compose file
failure message (there is additional output that i can provide but i will truncate it here as this seems the most relevant):
note: i have tried multiple models. best case scenario they return no response. worst case is it crashes like this. Would love to get this working on my server just for funsies. but im pretty sure the CPU is the limiting factor here. I know for a fact it does not have AVX so... thats a bad sign from the get go |
This is most likely caused by AVX support. You can compile local-ai on this machine to get a version optimized for it. |
Anyone else scouring through the issues for a solution, build it locally like bnusunny mentions above like this, but it does make it incredibly slow |
Yeah, without AVX, ggml will be slow on CPU. |
Does this work in the docker container? For me I got it to run locally with just |
|
LocalAI version:
quay.io/go-skynet/local-ai:latest
Environment, CPU architecture, OS, and Version:
IBM x3400 Server
with
Describe the bug
I'm new to localai and was trying to set-up the example "ChatGPT OSS Alternative" presented on localai-homepage. Link to example is: https://github.com/go-skynet/LocalAI/tree/master/examples/chatbot-ui
At first it looks like the localai-api is running fine, but sending any prompft using the chat-ui to the API causes crashing (see logs attached).
To Reproduce
Try this example:
https://github.com/go-skynet/LocalAI/tree/master/examples/chatbot-ui
This is my resulting docker-compose.yaml trying to adopt it:
Expected behavior
I've expected a working example with at least any output to the chat-gpt like prompt. But there's only "internal error" response popping up.
Logs
Log file from docker-container
Additional context
The text was updated successfully, but these errors were encountered: