view article Article Welcome Gemma 4: Frontier multimodal intelligence on device +5 3 days ago • 543
Marco-MoE Collection A suit of multilingual MoE models with highly-sparse architectures • 4 items • Updated 2 days ago • 8
Marco-MoE Collection A suit of multilingual MoE models with highly-sparse architectures • 4 items • Updated 2 days ago • 8
Marco-MoE Collection A suit of multilingual MoE models with highly-sparse architectures • 4 items • Updated 2 days ago • 8
Marco-MoE Collection A suit of multilingual MoE models with highly-sparse architectures • 4 items • Updated 2 days ago • 8
allenai/tulu-3-sft-personas-instruction-following Viewer • Updated Nov 21, 2024 • 30k • 5.36k • 63
Running on CPU Upgrade Featured 3.08k The Smol Training Playbook 📚 3.08k The secrets to building world-class LLMs