parallel-decoder-transformer / train_manifest.json
loganrobbins's picture
Publish PDT adapters + arXiv model card
4600161 verified
{
"agreement_threshold": 0.15,
"agreement_thresholds_file": "agreement_thresholds.json",
"best_eval_loss": 21.752999266554905,
"config_path": "/home/ubuntu/nstream-transformer/configs/gpt_oss_transfer_production.yaml",
"coverage_threshold": 0.4,
"dataset": "data/processed/pdt_10k_gpt41/kd_train.jsonl",
"eval_dataset": "data/processed/pdt_10k_gpt41/kd_validation.jsonl",
"git_dirty": true,
"git_sha": "d25d7dac8a57d6bed782e7251b657339341e33e0",
"global_step": 50000,
"notes_schema_version": "2.0",
"plan_hash_buckets": 65536,
"plan_hash_salt": "parallel-decoder-v1",
"plan_vocab_size": 65536,
"stages_file": "train_run_stages.json",
"wandb_run_name": "gpt-oss-8xH100-50000steps",
"wandb_run_url": "https://wandb.ai/ljrweb-self/parallel-decoder-transformer/runs/fmuea63a"
}