Waiting for updates W HEretic/Transformers to make this possible with "thinking" LFM base.
David Belton PRO
DavidAU
AI & ML interests
Application(s) of single/multiple LLMs in specialized use cases & automation tasks. LLM, Prompt , System Role and Parameter engineering VIA chat / API. 500+ LLMs graded.
Recent Activity
replied to
their
post
about 12 hours ago
Tiny but mighty: LFM 1.2B - 11 Distill / Fine tunes : Exceeding all benchmarks at 300-700+ T/S on GPU, 60+ T/S CPU.
Almost all exceed LFM 1.2B Benchmarks - which are already very impressive.
All benchmarks posted.
A specialized merge of multiple of these fine tunes by @nightmedia FAR exceeds the benchmarks set by the already impressive LFM.
(LFM2.5-1.2B-MEGABRAIN-Thinking-Polaris-ClaudeHOPUS-Deepseek-GLM)
Included are GLM 4.7 Flash, DeepSeek, Claude, Kimi V2 and other distill fine tunes.
Here is the collection ( Quants by MRadermarcher).
https://huggingface.co/collections/DavidAU/lfm-12b-sota-400-700-t-s-enhanced-fine-tunes-distills
updated
a collection
1 day ago
Thinking / Reasoning Models - Reg and MOEs.
updated
a collection
1 day ago
Qwen3 - 30B-A3B (128 experts) and higher
Organizations
None yet