view post Post 4255 We collaborated with Hugging Face to enable you to train MoE models 12× faster with 35% less VRAM via our new Triton kernels (no accuracy loss). 🤗Train gpt-oss locally on 12.8GB VRAM with our free notebooks: https://unsloth.ai/docs/new/faster-moe See translation 1 reply · 🔥 26 26 🤗 5 5 + Reply
Col-Bandit: Zero-Shot Query-Time Pruning for Late-Interaction Retrieval Paper • 2602.02827 • Published 10 days ago • 2
ibm-granite/granite-vision-3.3-2b-chart2csv-preview Image-Text-to-Text • 3B • Updated 3 days ago • 911 • 11
view post Post 3558 We created a tool-calling guide for local LLMs!Learn how to use any open model like Qwen3-Coder-Next and GLM-4.7-Flash for function calling.Guide: https://unsloth.ai/docs/basics/tool-calling-guide-for-local-llmsWe provide hands-on examples for: story writing, Python execution, terminal tool calls, maths and more. See translation 7 replies · ❤️ 17 17 + Reply
view post Post 3681 Qwen releases Qwen3-Coder-Next! 💜 Run the locally on 46GB RAM or less.Thhe model excels at agentic coding & local use. With 256K context, it delivers similar performance to models with 10-20× more active parameters.GGUF: unsloth/Qwen3-Coder-Next-GGUFGuide: https://unsloth.ai/docs/models/qwen3-coder-next See translation 9 replies · 🔥 13 13 🚀 5 5 🤯 2 2 + Reply
ibm-granite/granite-vision-3.3-2b-chart2csv-preview Image-Text-to-Text • 3B • Updated 3 days ago • 911 • 11
MILU: A Multi-task Indic Language Understanding Benchmark Paper • 2411.02538 • Published Nov 4, 2024 • 2
Influence Guided Sampling for Domain Adaptation of Text Retrievers Paper • 2601.21759 • Published 14 days ago • 1
ibm-granite/granite-vision-3.3-2b-chart2csv-preview Image-Text-to-Text • 3B • Updated 3 days ago • 911 • 11
Granite Experiments Collection Experimental projects under consideration for the Granite family. • 22 items • Updated 14 days ago • 15