C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & more LLMs
-
Updated
Apr 29, 2024 - C++
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & more LLMs
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
A Genshin Impact Book Question Answer Project supported by LLM
open-llms-next-web,一个类似于chatgpt-next-web的开源大型语言模型web演示,支持离线开源大模型和PEFT模型
LLM API Server , OpenAI 同时支持 ChatGLM3 ,Llama, Llama-3, Firefunction, Openfunctions ,BAAI/bge-m3 ,bge-large-zh-v1.5
This project accelerates local deployment of chatglm and vector inference using PyTorch compiled in C++, and includes an OpenAI API Mock script for quick setup of local speed testing services. This setup enhances performance and efficiency, ideal for high-performance applications and development testing.
An spoken English chatbot runs in realtime and offline based on LLM.
一个小玩具demo:使用LangChain调用本地ChatGLM3-6B模型实现的搜索引擎agent
Genshin Impact Character Instruction Models tuned by Lora on LLM
Add a description, image, and links to the chatglm3 topic page so that developers can more easily learn about it.
To associate your repository with the chatglm3 topic, visit your repo's landing page and select "manage topics."