Minimalist web-searching app with an AI assistant that runs directly from your browser. Uses Web-LLM, Ratchet-ML, Wllama and SearXNG. Demo: https://felladrin-minisearch.hf.space
-
Updated
May 18, 2024 - TypeScript
Minimalist web-searching app with an AI assistant that runs directly from your browser. Uses Web-LLM, Ratchet-ML, Wllama and SearXNG. Demo: https://felladrin-minisearch.hf.space
Design, conduct and analyze results of AI-powered surveys and experiments. Simulate social science and market research with large numbers of AI agents and LLMs.
Project Summa is an LLM-based text processing system built with Python and Django. It provides a modular and scalable solution for various text processing tasks.
A programming framework for agentic AI. Discord: https://aka.ms/autogen-dc. Roadmap: https://aka.ms/autogen-roadmap
Carbon Limiting Auto Tuning for Kubernetes
The open-source serverless GPU container runtime.
Local webui for Large Language Models. Supports the GGUF format. Inference LLMs with support for STT/TTS and function calling.
Empower Your Productivity with Local AI Assistants
Library to supercharge your use of large language models
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
Corpus2GPT: A project enabling users to train their own GPT models on diverse datasets, including local languages and various corpus types, using Keras and compatible with TensorFlow, PyTorch, or JAX backends for subsequent storage or sharing.
Leverage tensor parallelism techniques to run large language models in the CPU memory of edge devices.
LLM-PowerHouse: Unleash LLMs' potential through curated tutorials, best practices, and ready-to-use code for custom training and inferencing.
Run LLMs on weak devices or make powerful devices even more powerful by distributing the workload and dividing the RAM usage.
Template for RAG Applications
Miscellaneous codes and writings for MLOps
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
Add a description, image, and links to the llm-inference topic page so that developers can more easily learn about it.
To associate your repository with the llm-inference topic, visit your repo's landing page and select "manage topics."