Safetensors
llama

Covenant-72B

Covenant-72B is the largest permissionless collaboratively trained language model trained entirely from scratch at the 72 billion parameter scale.

It was trained with 20+ globally distributed participants coordinated via decentralized infrastructure on the Bittensor blockchain.

Covenant-72B


Training Details

Property Value
Model size 72B
Architecture LLaMA-style
Target token budget 1.1T
Compute participants 20+
Minimal compute per participant 8×B200 or equivalent
Dataset DCLM-baseline
Optimizer SparseLoCo (communication-efficient optimizer)

Performance on Benchmarks

All results are 0-shot acc-norm (%) unless noted.

Model Compute Environment / Permissions Size Tokens ARC-C ARC-E PIQA OpenBookQA HellaSwag Winogrande (acc) MMLU (acc)
Intellect-1 Internet / Whitelist 10B 1T 44.80 71.76 77.37 43.80 70.26 63.30 32.69
Psyche Consilience-7Y9 Internet / Whitelist 40B 1.2T 31.14 55.77 76.12 35.20 63.67 56.99 24.23
Covenant-72B (Checkpoint-Two) Internet / Permissionless 72B 420B 53.84 77.74 80.58 44.60 77.08 71.43 47.49
Covenant-72B (base) Internet / Permissionless 72B 1.1T 56.48 79.76 80.90 44.80 78.07 73.24 61.00
LLM360 K2 ckpt_108 Centralized Cluster 65B 420B 45.73 70.54 80.90 43.20 78.23 71.90 50.01
LLM360 K2 Stage 1 Centralized Cluster 65B 1.4T 53.75 75.97 82.54 48.00 82.86 76.40 65.51
LLaMA-2-7B Centralized Cluster 7B 2T 45.05 73.82 78.73 44.20 76.18 69.38 41.73
LLaMA-2-70B Centralized Cluster 70B 2T 57.42 79.55 82.59 49.40 84.34 80.43 65.63

Downloads last month
81
Safetensors
Model size
73B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train 1Covenant/Covenant-72B