Change8

Unsloth

AI & LLMs

Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek, Qwen, Llama, Gemma, TTS 2x faster with 70% less VRAM.

Latest: December-202514 releases3 breaking changesView on GitHub →

Release History

December-2025Breaking11 fixes13 features
Dec 18, 2025

This release introduces massive performance gains with 3x faster training via new Triton kernels and enables 500K context length fine-tuning. It also adds support for Transformers v5, preliminary multi-GPU training, and several new model guides.

November-202515 fixes11 features
Nov 25, 2025

This release introduces major performance enhancements with FP8 Reinforcement Learning support and significant VRAM reductions across the board. It also adds support for new models like DeepSeek-OCR and Qwen3-VL, alongside improved Docker integration.

October-202524 fixes14 features
Oct 27, 2025

This release introduces major platform support, including Docker images and Blackwell/DGX hardware compatibility, alongside significant new features like Quantization-Aware Training (QAT) and extensive RL environment utilities.

September-2025-v3Breaking4 fixes4 features
Sep 26, 2025

This release introduces significant performance gains and new capabilities for gpt-oss Reinforcement Learning, alongside support for new models like DeepSeek-V3.1-Terminus and Magistral 1.2. Several bug fixes were implemented, including resolving issues with BERT and QAT + LoRA fast path.

September-2025-v212 fixes8 features
Sep 16, 2025

This release introduces major performance enhancements and new capabilities for Vision models in Reinforcement Learning (RL), alongside the new 'Standby' feature for memory-efficient training. Numerous bug fixes and improvements were also integrated across various components, including Intel/ROCm support and serialization workflows.

August-2025-v29 fixes7 features
Aug 28, 2025

This release introduces Unsloth Flex Attention for gpt-oss training, drastically improving context length, VRAM efficiency, and speed. Numerous bug fixes and support for new models/features like QAT + LoRA are also included.

August-202519 fixes8 features
Aug 8, 2025

This release introduces broad support for the new gpt-oss model, enabling low-VRAM fine-tuning, alongside significant algorithmic updates that improve performance across all models. It also adds support for Qwen3 models and expands compatibility to include newer NVIDIA hardware like RTX 50 series and Blackwell GPUs.

July-202518 fixes7 features
Jul 10, 2025

This release focuses heavily on stability, VRAM reduction (10-25% less), and broad model compatibility, including full fixes for Gemma 3N Vision and support for new models like Devstral 1.1 and MedGemma.

June-2025Breaking23 fixes16 features
Jun 26, 2025

This release introduces major new capabilities including support for multimodal Gemma 3n models and Text-to-Speech fine-tuning, alongside new quantization methods (Dynamic 2.0 GGUFs) and support for DeepSeek-R1-0528 and Magistral-24B.

May-202512 fixes5 features
May 2, 2025

This release introduces official support for Qwen3 models, including fine-tuning capabilities for the 30B MoE variant. Numerous bug fixes address compatibility issues, quantization errors, and improve overall stability.

2025-037 fixes15 features
Mar 14, 2025

The March release introduces full support for finetuning Gemma 3 models and significantly expands model compatibility, including Mixtral and vision models, alongside preliminary support for 8bit and full finetuning. This version also brings Windows support and removes the compilation requirement for GGUF exports.

2025-02-v25 fixes3 features
Feb 20, 2025

This release introduces GRPO, achieving up to 90% memory reduction during training, alongside various bug fixes and updates to support Llama 3.1 8B training.

2025-029 fixes7 features
Feb 6, 2025

This release introduces major support for GRPO training, enabling LoRA/QLoRA for GRPO across various models, and integrates fast inference via vLLM for significant throughput gains. Numerous bug fixes address issues with Gemma 2, Mistral mapping, and general stability.

2025-0114 fixes6 features
Jan 10, 2025

This release introduces full support for the Phi-4 model, including fixes for tokenization and chat templates, alongside significant bug fixes for gradient accumulation, vision models, and performance regressions. It also brings Windows support and performance improvements via updated Xformers.