Load larger models by offloading model layers to both GPU and CPU
-
Updated
Jul 28, 2023 - Jupyter Notebook
Load larger models by offloading model layers to both GPU and CPU
Ask LLaMa about image in your clipboard
Auto Complete anything using a gguf model
Use your open source local model from the terminal
Lightweight implementation of the OpenAI open API on top of local models
A chatbot with the ability to vocally respond (TTS) using llama
Presentation on Artificial Intelligence for the Free Drawing and Print Graphics class of the Muthesius Academy of Art.
Repo to download, save and run quantised LLM models using Llama.cpp and benchmark the results (private use)
Llama-2 on apple mac using gpu
UnOfficial Gradio Repo for ICML 2024 paper "Executable Code Actions Elicit Better LLM Agents" by Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhang, Yunzhu Li, Hao Peng, Heng Ji.
A custom framework for easy use of LLMs, VLMs, etc. supporting various modes and settings via web-ui
AgentX is an Open-source library that help people use LLMs on their own computers or help them to serve LLMs as easy as possible that support multi-backends like PyTorch, llama.cpp, Ollama and EasyDeL
PowerShell automation to download large language models (LLMs) from Git repositories and quantize them with llama.cpp into the GGUF format.
YouTube API implementation with Meta's Llama 2 to analyze comments and sentiments
Add a description, image, and links to the llama-cpp topic page so that developers can more easily learn about it.
To associate your repository with the llama-cpp topic, visit your repo's landing page and select "manage topics."