How bad is the precision loss on the Q2_K_XL quants? I can run that with full GPU offload but I usually dont run lower than Q4, broadly.
Ben Kelly PRO
YellowjacketGames
AI & ML interests
None yet
Recent Activity
replied to
danielhanchen's
post
about 17 hours ago
You can now run MiniMax-2.5 locally! 🚀
At 230B parameters, MiniMax-2.5 is the strongest LLM under 700B params, delivering SOTA agentic coding & chat.
Run Dynamic 3/4-bit on a 128GB Mac for 20 tokens/s.
Guide: https://unsloth.ai/docs/models/minimax-2.5
GGUF: https://huggingface.co/unsloth/MiniMax-M2.5-GGUF
liked
a model
5 days ago
PrimeIntellect/INTELLECT-3
updated
a collection
6 days ago
[mixed] Chess x AI