Zixi "Oz" Li PRO
OzTianlu
AI & ML interests
My research focuses on deep reasoning with small language models, Transformer architecture innovation, and knowledge distillation for efficient alignment and transfer.
Recent Activity
updated
a model 16 minutes ago
NoesisLab/Spartacus-1B-Instruct reacted
to
their post with π€ 31 minutes ago
π‘οΈ Meet Spartacus-1B: Shattering the Memory Wall with True O(1) Inference! π
https://huggingface.co/NoesisLab/Spartacus-1B-Instruct
https://huggingface.co/spaces/NoesisLab/ChatSpartacus
At NoesisLab, we've entirely ripped out Softmax Attention and replaced it with Causal Monoid State Compression.
Say hello to Spartacus-1B-Instruct (1.3B) π‘οΈ.
Instead of maintaining a massive, ever-growing list of past tokens, Spartacus compresses its entire causal history into a fixed-size state matrix per head. The result?
β‘ True O(1) Inference: Memory footprint and generation time per token remain absolutely constant, whether you are on token 10 or token 100,000.
π§ Explicit Causality: We threw away RoPE and attention masks. The model learns when to forget using dynamic, content-aware vector decay.
π₯ Blazing Fast Training: Full hardware utilization via our custom Triton-accelerated JIT parallel prefix scan.
π Zero-Shot Benchmarks that Hit Hard:
O(1) architectures usually sacrifice zero-shot accuracy. Not Spartacus. It is punching way above its weight class, beating established sub-quadratic models (like Mamba-1.4B and RWKV-6-1.6B):
π ARC-Challenge: 0.3063 (vs Mamba 0.284)
π ARC-Easy: 0.5518
π PIQA: 0.6915 reacted
to
their post with π₯ 37 minutes ago
π‘οΈ Meet Spartacus-1B: Shattering the Memory Wall with True O(1) Inference! π
https://huggingface.co/NoesisLab/Spartacus-1B-Instruct
https://huggingface.co/spaces/NoesisLab/ChatSpartacus
At NoesisLab, we've entirely ripped out Softmax Attention and replaced it with Causal Monoid State Compression.
Say hello to Spartacus-1B-Instruct (1.3B) π‘οΈ.
Instead of maintaining a massive, ever-growing list of past tokens, Spartacus compresses its entire causal history into a fixed-size state matrix per head. The result?
β‘ True O(1) Inference: Memory footprint and generation time per token remain absolutely constant, whether you are on token 10 or token 100,000.
π§ Explicit Causality: We threw away RoPE and attention masks. The model learns when to forget using dynamic, content-aware vector decay.
π₯ Blazing Fast Training: Full hardware utilization via our custom Triton-accelerated JIT parallel prefix scan.
π Zero-Shot Benchmarks that Hit Hard:
O(1) architectures usually sacrifice zero-shot accuracy. Not Spartacus. It is punching way above its weight class, beating established sub-quadratic models (like Mamba-1.4B and RWKV-6-1.6B):
π ARC-Challenge: 0.3063 (vs Mamba 0.284)
π ARC-Easy: 0.5518
π PIQA: 0.6915