Skip to content

Serving open source models of your choice in as a docker container using llama-cpp-python's OpenAI compatible server

Notifications You must be signed in to change notification settings

darshpanchal/llm-server

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

llm-server

Simple Docker script to create an OpenAI compatible server using llama-cpp-python on port 8000.

Usage

Build the Docker Container

docker build -t llm-server .

Run the Docker container

docker run -d --name llmserver -p 8000:8000 -v $PWD/config:/home/config -v $PWD/models:/home/models llm-server

You can change config.json to change port or model information. Supports multiple models, check llama-cpp-python documentation.

Note: After updating config.json, you need to restart the container for changes to take effect.

About

Serving open source models of your choice in as a docker container using llama-cpp-python's OpenAI compatible server

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published