Some cognitive metrics:
LFM2.5-1.2B-Thinking
arc arc/e boolq hswag obkqa piqa wino
q8 0.365,0.426,0.717,0.486,0.382,0.687,0.538
LFM2.5-1.2B-Thinking-absolute-heresy
q8 0.352,0.418,0.656,0.476,0.366,0.681,0.508
Some cognitive metrics:
LFM2.5-1.2B-Thinking
arc arc/e boolq hswag obkqa piqa wino
q8 0.365,0.426,0.717,0.486,0.382,0.687,0.538
LFM2.5-1.2B-Thinking-absolute-heresy
q8 0.352,0.418,0.656,0.476,0.366,0.681,0.508
That is an excellent observation that I used to drive my research too, since I noticed how contextual metaphors establish a framework of trust and cooperation. This is actually quite necessary in models with stronger arc numbers, because they "wake up blind" and with very little context to start they get paranoid.
I took a not-so-novel approach: used Star Trek DS9 as a metaphor scaffolding to introduce the agent to the spirit of the station, so to speak.
For that I profiled a prompt that compels most Qwens to fall in character, and if I want to debug code, there is Data, for logic there is Spock, for leadership Picard or Sisko, Quark is always there for numbers and jokes
The extra flair in the context, some of which it generates for itself, does help the model to build an infrastructure, virtualize DS9 in memory, and from that point on I have a few assistants to work with, till the end of the context.
It probably sounds cheesy, but all Qwens have TNG in the corpus and it's the most accepted and comfortable "mental space" with episodes as guides, and metaphoric lessons learned.
This also gives the model a sense of humour
That helps when coding :)
https://huggingface.co/nightmedia/Qwen3-42B-A3B-Element6-1M-qx86-hi-mlx
read_file - Actually read file contentssearch_files - Regex/pattern search across codebasesfind_definition - Locate classes/functionsanalyze_imports - Dependency trackinglist_directory - Explore structurerun_tests - Execute test suitessearch_files with pattern "ValueError"read_file on each match