Call many AIs from a single API.
-
Updated
Mar 28, 2024 - Go
Call many AIs from a single API.
Evaluate open-source language models on Agent, formatted output, command following, long text, multilingual, coding, and custom task capabilities. 开源语言模型在Agent,格式化输出,指令追随,长文本,多语言,代码,自定义任务的能力基准测试。
EchoSight is a tool that helps visually impaired individuals by audibly describing images taken with a Raspberry Pi Camera or inputted via image path or URL across different operating systems.
MLOps library for LLM deployment w/ the vLLM engine on RunPod's infra.
An simple implementation of Unet because all the implementations i've seen are wayy tooo complicated.
This repository has a lot of LLM projects done. It is the best place to start learning LLM.
A discord bot which can call LLMs using either Hugging Face or vLLM on Windows platform. Combined with function calling.
Run code inference-only benchmarks quickly using vLLM
Preserving entities through the integration of knowledge graphs, Llama 2, vLLM, and LangChain.
Embedding based semantic search app for poetry [App and EDA notebooks]
AI-Learning-Platform, a LLM-RAG pipeline which behaves like a guide and able to solve doubts. Deployed on-premise IBM ppc64le architecture. vLLM for model inference & Qdrant with Langchain for RAG Pipeline. Server written in django, postgres & cassandra as the sql & nosql databases.
Dockerized LLM inference server with constrained output (JSON mode), built on top of vLLM and outlines. Faster, cheaper and without rate limits. Compare the quality and latency to your current LLM API provider.
Add a description, image, and links to the vllm topic page so that developers can more easily learn about it.
To associate your repository with the vllm topic, visit your repo's landing page and select "manage topics."