Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Methods to load the local downloaded LLM model #83

Open
lawyinking opened this issue Jan 10, 2024 · 2 comments
Open

Methods to load the local downloaded LLM model #83

lawyinking opened this issue Jan 10, 2024 · 2 comments

Comments

@lawyinking
Copy link

in Lida, is it possible to load the model downloaded on a MacBook M2? say the Zephyr 7b or Magicoder 7b, I downloaded the models on a Desktop folder, then I would like to load it using some simple method like llama-cpp-python, or Langchain's LlamaCpp, but not loading it using methods that Lida suggested, like llmx load model from hugging face, vllm server, OpenAI etc.

Thanks!

@kishoretvk
Copy link

anything on this ?
can we use quantized local models ?

@victordibia
Copy link
Collaborator

What quantized models are you interested in using?
In general, as long as you can spin up an openai compliant webserver endpoint from your model, you can integrate it into lida.

Most local models have tools that can spin up an openai compliant api, even llama.cpp.

Once you spin up the openai compliant api, you can simply just use it with lida (see the llmx api docs)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants