gemma
Here are 90 public repositories matching this topic...
🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
-
Updated
May 18, 2024 - C++
Finetune Llama 3, Mistral & Gemma LLMs 2-5x faster with 80% less memory
-
Updated
May 18, 2024 - Python
Documentation for Google's Gen AI site - including the Gemini API and Gemma
-
Updated
May 18, 2024 - Jupyter Notebook
Firefly: 大模型训练工具,支持训练Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
-
Updated
May 9, 2024 - Python
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
-
Updated
May 17, 2024 - Python
Fully-featured, beautiful web interface for Ollama LLMs - built with NextJS. Deploy with a single click.
-
Updated
Apr 30, 2024 - TypeScript
🚀🚀🚀A collection of some awesome public projects about Large Language Model, Vision Foundation Model and AI Generated Content.
-
Updated
May 9, 2024
A snappy, keyboard-centric terminal user interface for interacting with large language models. Chat with ChatGPT, Claude, Llama 3, Phi 3, Mistral, Gemma and more.
-
Updated
May 14, 2024 - Python
Collecting data for Building Lucknow's first LLM
-
Updated
May 9, 2024 - Jupyter Notebook
Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).
-
Updated
Mar 15, 2024 - C++
Unofficial PyTorch/🤗Transformers(Gemma/Llama3) implementation of Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention
-
Updated
Apr 23, 2024 - Python
🏗️ Fine-tune, build, and deploy open-source LLMs easily!
-
Updated
May 17, 2024 - Go
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
-
Updated
May 18, 2024 - Python
Examples of RAG using Llamaindex with local LLMs - Gemma, Mixtral 8x7B, Llama 2, Mistral 7B, Orca 2, Phi-2, Neural 7B
-
Updated
Feb 25, 2024 - Jupyter Notebook
Train GEMMA on TPU/GPU! (Codebase for training Gemma-Ko Series)
-
Updated
Mar 2, 2024 - Python
PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"
-
Updated
May 18, 2024 - Python
Improve this page
Add a description, image, and links to the gemma topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the gemma topic, visit your repo's landing page and select "manage topics."