July-2025
📦 unslothView on GitHub →
✨ 7 features🐛 18 fixes🔧 20 symbols
Summary
This release focuses heavily on stability, VRAM reduction (10-25% less), and broad model compatibility, including full fixes for Gemma 3N Vision and support for new models like Devstral 1.1 and MedGemma.
Migration Steps
- Update Unsloth using: `pip install --upgrade --force-reinstall --no-deps --no-cache-dir unsloth unsloth_zoo`
- For users of Gemma 3N, re-download all model checkpoints (Unsloth will handle this automatically).
✨ New Features
- Gemma 3N Vision support is now functional.
- Gemma 3 text and vision performance significantly improved on T4 hardware.
- Achieved 10 to 25% reduction in VRAM consumption across all models.
- Added support for Devstral 1.1 and MedGemma 27B, 4B models, including vision capabilities.
- Enabled Unsloth support for AMD GPUs via vLLM.
- Added Intel GPU support with vLLM integration.
- Falcon H1 finetuning is now supported.
🐛 Bug Fixes
- Fixed issues with Gemma 3 text and vision training losses (now 1 to 2 instead of 6 to 7).
- Resolved downloads getting stuck between 90% and 95%.
- Fixed compatibility issues for Qwen 2.5, Qwen 2, and GLM models.
- Resolved GRPO compatibility with the latest main TRL branch.
- Ensured compatibility with the latest main branches of TRL, PEFT, and Transformers.
- Fixed an issue where transformers were being forcibly upgraded.
- Fixed LoftQ implementation when used with FastBaseModel.
- Fixed LoftQ None config issue for FastBaseModel.
- Converted torch.bfloat16, torch.float16, etc., to vLLM valid dtypes.
- Fixed quantized model parameter count method.
- Fixed UnslothTrainingArguments not patching trl.Config correctly.
- Fixed unnecessary warning when using transformers >= 4.53.0.
- Silently skipped Falcon H1 import if transformers_version is less than 4.53.0.
- Dynamically adjusted get_per_token_logps for TRL main upgrade.
- Fixed casual mask implementation.
- Explicitly checked for xformers existence before using it for attention.
- Added fallback check for feed_forward if mlp does not exist in the layer module for Falcon H1.
- Fixed device placement by moving inputs to the correct devices.
🔧 Affected Symbols
Gemma 3NGemma 3Qwen 2.5Qwen 2GLMGRPOTRLPEFTTransformersFalcon H1Devstral 1.1MedGemma 27BMedGemma 4BLoftQFastBaseModelUnslothTrainingArgumentstrl.Configtorch.bfloat16torch.float16get_per_token_logps