12x Faster MoE Training + Embedding support! #4020
danielhanchen
announced in
Announcements
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Our first release of 2026! This year we’ve got a lot of exciting things coming and to kick things off, we’re introducing faster MoE training, embedding model support, and ultra long context for Reinforcement Learning. We’ll also be launching our brand new UI very soon.
We’d like to thank all of you for 50K stars on GitHub! ⭐
We’ve also added support for many new models that you can now run and fine-tune locally, including DeepSeek-OCR 2, GLM-4.7-Flash, Kimi-2.5, and more.
🚀 Faster MoE training
You can now train MoE models 12× faster with 35% less VRAM and 6x longer context via our new Triton and math kernels (no accuracy loss). gpt-oss-20b works on 12.8GB VRAM. Qwen3-30B-A3B (16-bit LoRA) uses 63GB.
Unsloth supports fast training for gpt-oss, Qwen3 (30B, 235B, VL, Coder), DeepSeek R1/V3 arch and GLM (4.7, Flash) models.
Faster MoE Blog
🔎 Embedding models now train 2× faster
We collaborated with Hugging Face to enable 1.8-3.3x faster embedding, BERT and classifier model training with 20% less VRAM, 2x longer context & no accuracy loss vs. FA2 setups.
Embedding model Blog
💡 Ultra Long Context RL is here
We’re introducing new batching algorithms to enable ~7x longer context (can be more than 12x) RL training with no accuracy or speed degradation vs. other optimized setups that use FA3, kernels & chunked losses.
Unsloth now trains gpt-oss QLoRA with 380K context on a single 192GB NVIDIA B200 GPU
Long Context RL Blog
🔮 New models
🎉 Extra Updates
trl==0.27.1andtransformers==5.1.0are supported well - previous coverage was 30% of all our 120 notebooks, but now we have >80% coverage - we plan to make it 100% over the next few days.📖 New Guides
Tip
Update Unsloth via
pip install --upgrade --force-reinstall --no-cache-dir --no-deps unsloth unsloth_zooIf you want PyTorch 2.9:
pip install --upgrade unsloth unsloth_zooFebruary is shaping up to be an amazing month for LLM releases, and we hope you’re just as excited as we are. 😊
What's Changed
Unsloth Zoo Changes
FastLanguageModel.for_inference()whenmodel.eval()is called #392, transformers 5 by @electroglyph in fix for #392, transformers 5 unsloth-zoo#393New Contributors
Full Changelog: December-2025...February-2026
This discussion was created from the release 12x Faster MoE Training + Embedding support!.
Beta Was this translation helpful? Give feedback.
All reactions