OpenResearcher/OpenResearcher-30B-A3B Feature Extraction • 32B • Updated about 22 hours ago • 10.2k • 53
view post Post 5124 We collaborated with Hugging Face to enable you to train MoE models 12× faster with 35% less VRAM via our new Triton kernels (no accuracy loss). 🤗Train gpt-oss locally on 12.8GB VRAM with our free notebooks: https://unsloth.ai/docs/new/faster-moe See translation 1 reply · 🔥 29 29 🤗 5 5 + Reply
view post Post 4519 I just released NovaSR, a tiny 52kb audio upsampler that can enhance 3600 seconds of muffled 16khz audio in to clearer 48khz audio in just 1 second!NovaSR can- Enhance TTS model quality.- Restore poor quality datasets.- Work on any device(just 52kb which is smaller than a 3 second audio file!)Model: YatharthS/NovaSRSpace to try it: YatharthS/NovaSRGithub repo: https://github.com/ysharma3501/NovaSR See translation 5 replies · 🚀 18 18 🔥 6 6 + Reply
view post Post 365 Agentic capability is the new battleground🔥LongCat-Flash-Thinking-2601, the latest reasoning model from Meituan- LongCat✨ MoE - 560B total / 27B active✨ MIT license ✨ Agentic tool use✨ Multi-environment RL✨ Parallel + iterative reasoning meituan-longcat/LongCat-Flash-Thinking-2601 See translation 🔥 2 2 + Reply
view post Post 2858 You can now do reinforcement learning training with 7× longer context and no accuracy loss, via our new batching algorithms.Long reasoning chains in RL are costly, but now we enable you to train gpt-oss with GRPO & reach 380K context on a 192GB GPU.Blog: https://unsloth.ai/docs/new/grpo-long-context See translation 🔥 8 8 ❤️ 4 4 🚀 3 3 + Reply