An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
-
Updated
May 17, 2024 - Python
An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
Various LMs/LLMs below 3B parameters (for now) trained using SFT (Supervised Fine Tuning) for several downstream tasks
InternLM-XComposer2 is a groundbreaking vision-language large model (VLLM) excelling in free-form text-image composition and comprehension.
[ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".
This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & Vertical Distillation of LLMs.
Binary classification of pathological heartbeats from ECG signals using 1D CNNs in PyTorch
Knowledge Verification to Nip Hallucination in the Bud
Knowledge Verification to Nip Hallucination in the Bud
Finetuning Google's Gemma Model for Translating Natural Language into SQL
Finetune Mistral 7b v1.0 on custom dataset
Fine tune Large Language Model on Mathematic dataset
A LLM challenge to (i) fine-tune pre-trained HuggingFace transformer model to build a Code Generation language model, and (ii) build a retrieval-augmented generation (RAG) application using LangChain
This project streamlines the fine-tuning process, enabling you to leverage Llama-2's capabilities for your own projects.
The offical realization of InstructERC
Aligning Large Language Models with Human: A Survey
🦙 Llama 2 7B fine-tuned to revive Rick
开源SFT数据集整理,随时补充
Add a description, image, and links to the supervised-finetuning topic page so that developers can more easily learn about it.
To associate your repository with the supervised-finetuning topic, visit your repo's landing page and select "manage topics."