An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
-
Updated
May 17, 2024 - Python
An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
InternLM-XComposer2 is a groundbreaking vision-language large model (VLLM) excelling in free-form text-image composition and comprehension.
开源SFT数据集整理,随时补充
Aligning Large Language Models with Human: A Survey
This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & Vertical Distillation of LLMs.
The offical realization of InstructERC
[ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".
Finetuning Google's Gemma Model for Translating Natural Language into SQL
Knowledge Verification to Nip Hallucination in the Bud
Various LMs/LLMs below 3B parameters (for now) trained using SFT (Supervised Fine Tuning) for several downstream tasks
This project streamlines the fine-tuning process, enabling you to leverage Llama-2's capabilities for your own projects.
Binary classification of pathological heartbeats from ECG signals using 1D CNNs in PyTorch
🦙 Llama 2 7B fine-tuned to revive Rick
Fine tune Large Language Model on Mathematic dataset
Knowledge Verification to Nip Hallucination in the Bud
Finetune Mistral 7b v1.0 on custom dataset
A LLM challenge to (i) fine-tune pre-trained HuggingFace transformer model to build a Code Generation language model, and (ii) build a retrieval-augmented generation (RAG) application using LangChain
Add a description, image, and links to the supervised-finetuning topic page so that developers can more easily learn about it.
To associate your repository with the supervised-finetuning topic, visit your repo's landing page and select "manage topics."