Covenant72B

Covenant72B is the largest permissionless collaboratively trained language model trained entirely from scratch at the 72 billion parameter scale.

It is being trained with 20+ globally distributed participants coordinated via decentralized infrastructure on the Bittensor blockchain.

Checkpoint-One marks the first release, corresponding to 200 billion tokens processed. Model files are available in the Checkpoint-One branch. Future checkpoints will be updated here.

Checkpoint One


Training Details

Property Value
Model size 72B
Architecture LLaMA-style
Target token budget 1.2T (210B for current checkpoint)
Compute participants 20+
Minimal compute per participant 8×B200 or equivalent
Dataset DCLM-baseline
Optimizer SparseLoCo (communication-efficient optimizer)

Performance on Benchmarks

All results are 0-shot acc-norm (%)

Model Compute Environment / Permissions Size Tokens ARC-C ARC-E PIQA OpenBookQA HellaSwag Winogrande MMLU
Intellect-1 Over the internet / White List 10B 1T 44.8 71.6 77.7 43.6 70.5 63.1 32.7
Psyche Consilience-7Y9 Over the internet / White List 40B 1.2T 31.1 55.8 76.1 34.8 63.7 57.0 24.2
Covenant72B – Checkpoint One Over the internet / Permissionless 70B 210B 46.2 72.6 79.2 43.0 73.5 70.3 38.0
K2 Checkpoint 54 Centralized Cluster 65B 210B 41.8 69.5 80.1 42.4 74.9 68.9 33.7

For more details, refer to Checkpoint One on Templar Research.

Downloads last month
188
Safetensors
Model size
73B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for diagonalge/Covanent-Base-Non-Chat

Adapters
8 models

Dataset used to train diagonalge/Covanent-Base-Non-Chat