Releases
v0.5.0
v0.5.0: Agent Tuning, Unsloth Integration
Congratulations on 10k stars π Make LLM fine-tuning easier and faster together with LLaMA-Factory β¨
New features
Support agent tuning for most models, you can fine-tune any LLMs with --dataset glaive_toolcall
for tool using #2226
Support function calling in both API and Web mode with fine-tuned models, same as the OpenAI's format
LLaMA Factory π€ Unsloth , enjoy 170% LoRA training speed with --use_unsloth
, see benchmarking here
Supports fine-tuning models on MPS device #2090
New models
Base models
Phi-2 (2.7B)
InternLM2 (7B/20B)
SOLAR-10.7B
DeepseekMoE-16B-Base
XVERSE-65B-2
Instruct/Chat models
InternLM2-Chat (7B/20B)
SOLAR-10.7B-Instruct
DeepseekMoE-16B-Chat
Yuan (2B/51B/102B)
New datasets
Supervised fine-tuning datasets
deepctrl dataset
Glaive function calling dataset v2
Core updates
Refactor data engine: clearer dataset alignment, easier templating and tool formatting
Refactor saving logic for models with value head #1789
Use ruff code formatter for stylish code
Bug fix
You canβt perform that action at this time.