-
Notifications
You must be signed in to change notification settings - Fork 43
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
OpenAI compatible chat completions endpoint #28
Comments
You can use server provided by llama.cpp https://github.com/kaleid-liner/llama.cpp/tree/master/examples%2Fserver |
Spent 30 minutes yesterday, dropped my attempts after failing to build it... Would be great to have some instruction |
Btw, recent llama.cpp has different way of interacting with the server, it's first class citizen now (not part of the examples) -
|
After successful running run_pipeline.py, you can just run |
I get some dependency error
|
Only after we have merged the latest llama.cpp, see #24 |
I think this is not a issue of T-MAC, and the problem is described straightforward. You need to provide openssl. You can try to solve it by searching in llama.cpp. |
That was like problem number 10 after I spent whole evening building and running T-MAC :) Got exhausted at this point and had no courage to proceed with tinkering... I am no C dev, would be happy to return to T-MAC when someone experienced fixes the server feature and provides clear instructions. Thx for an interesting project, BTW |
Sorry for the inconvenience. Currently we have no good solution to simplify the build process. |
Would be great to have an easy way to run OpenAI endpoint on localhost and have the ability to interface with the model via HTTP API - e.g. use with any of the chat bot UI options
The text was updated successfully, but these errors were encountered: