AI & ML interests

None defined yet.

Recent Activity

danielhanchenΒ 
posted an update 1 day ago
danielhanchenΒ 
posted an update 7 days ago
view post
Post
2538
You can now fine-tune embedding models in our free Unsloth notebook! πŸ€—

Fine-tuning embedding models improves retrieval & RAG by aligning vectors to your domain-specific notion of similarity, improving search, clustering, and recommendations on your data.

⭐ Blog + Notebooks: https://unsloth.ai/docs/new/embedding-finetuning

Unsloth trains embedding models 1.8-3.3x faster with 20% less VRAM, 2x longer context & no accuracy loss vs. FA2 setups.

We'd like to thank Hugging Face and Unsloth contributor: electroglyph for making this possible!
  • 2 replies
Β·
danielhanchenΒ 
posted an update 9 days ago
danielhanchenΒ 
posted an update 14 days ago
view post
Post
2804
You can now do reinforcement learning training with 7Γ— longer context and no accuracy loss, via our new batching algorithms.

Long reasoning chains in RL are costly, but now we enable you to train gpt-oss with GRPO & reach 380K context on a 192GB GPU.

Blog: https://unsloth.ai/docs/new/grpo-long-context
danielhanchenΒ 
posted an update 29 days ago
danielhanchenΒ 
posted an update about 1 month ago
danielhanchenΒ 
posted an update about 1 month ago
danielhanchenΒ 
posted an update about 2 months ago
danielhanchenΒ 
posted an update about 2 months ago
danielhanchenΒ 
posted an update about 2 months ago
view post
Post
3851
Mistral's new Ministral 3 models can now be Run & Fine-tuned locally! (16GB RAM)
Ministral 3 have vision support and the best-in-class performance for their sizes.
14B Instruct GGUF: unsloth/Ministral-3-14B-Instruct-2512-GGUF
14B Reasoning GGUF: unsloth/Ministral-3-14B-Reasoning-2512-GGUF

🐱 Step-by-step Guide: https://docs.unsloth.ai/new/ministral-3
All GGUFs, BnB, FP8 etc. variants uploads: https://huggingface.co/collections/unsloth/ministral-3
Β·
danielhanchenΒ 
posted an update 2 months ago
JingzeShiΒ 
posted an update 3 months ago
danielhanchenΒ 
posted an update 3 months ago
view post
Post
4448
You can now run Kimi K2 Thinking locally with our Dynamic 1-bit GGUFs: unsloth/Kimi-K2-Thinking-GGUF

We shrank the 1T model to 245GB (-62%) & retained ~85% of accuracy on Aider Polyglot. Run on >247GB RAM for fast inference.

We also collaborated with the Moonshot AI Kimi team on a system prompt fix! πŸ₯°

Guide + fix details: https://docs.unsloth.ai/models/kimi-k2-thinking-how-to-run-locally