allenai/Olmo-3-7B-Think-SFT
Text Generation
•
7B
•
Updated
•
1.75k
•
5
All artifacts for post-training Olmo 3. Datasets follow the model that resulted from training on them.
Note Our 7B Think model lineage.
Note Our 32B Think SFT dataset is slightly further filtered than the 7B version above.
Note The new Python code mix we created for our thinking models. We took a subset of this for our full Think SFT Mix.
Note Our RL Zero 7B models with intermediate checkpoints per domain.
Note Our RL Zero datasets.