-
-
Notifications
You must be signed in to change notification settings - Fork 2.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: direct llama.cpp integration #1483
Comments
Just a quick follow-up to say it seems to work fine:
and it seems to be calling the OAI-like API endpoint on the llama.cpp server fine. It wasn't that clear I needed to add the The only difference I can see is there is no little "information" icon like there was with Ollama models, but it does seem to be calling the
I'll report back if I can see any other major differences, but otherwise 👍 |
I've used this quite a bit with llama.cpp server now and the only problem I've come across is pressing the stop button doesn't actually disconnect/stop the generation. This was a problem with the Ollama server and was fixed AFAIK: It would be helpful if this could be added to the OpenAI API code too, as otherwise the only way currently to stop runaway LLMs is to Control-C the running server and restart it. |
Another thing that might be helpful would be to add an option to hide the "Modelfiles" and "Prompts" menu options in the left, as these aren't able to be used with the OpenAI API and just add clutter. |
@jukofyork I'll start working on this feature after #665, we should strive to keep all the core features. |
Small update: Stop generation button is still an issue |
@DenisSergeevitch that is unrelated to the issue being discussed here. Let's keep discussion of the stop generation function here: |
No description provided.
The text was updated successfully, but these errors were encountered: