[2025-10-30 00:38:06,468] [DEBUG] [axolotl.utils.config.resolve_dtype:66] [PID:3449] bf16 support detected, enabling for this configuration. config.json: 0%| | 0.00/601 [00:00" }, "streaming_multipack_buffer_size": 10000, "strict": false, "tensor_parallel_size": 1, "tf32": true, "tiled_mlp_use_original_mlp": true, "tokenizer_config": "mistralai/Mistral-7B-Instruct-v0.3", "tokenizer_save_jinja_files": true, "tokenizer_type": "LlamaTokenizer", "torch_dtype": "torch.bfloat16", "train_on_inputs": false, "trl": { "log_completions": false, "mask_truncated_completions": false, "ref_model_mixup_alpha": 0.9, "ref_model_sync_steps": 64, "scale_rewards": true, "sync_ref_model": false, "use_vllm": false, "vllm_server_host": "0.0.0.0", "vllm_server_port": 8000 }, "trust_remote_code": true, "type_of_model": "MistralForCausalLM", "use_otel_metrics": false, "use_ray": false, "use_wandb": true, "val_set_size": 0.05, "vllm": { "device": "auto", "dtype": "auto", "gpu_memory_utilization": 0.9, "host": "0.0.0.0", "port": 8000 }, "wandb_log_model": "", "wandb_name": "mistral-7b-legal-v0.3", "wandb_project": "mistral-legal-french", "wandb_watch": "", "warmup_steps": 100, "weight_decay": 0.01, "world_size": 1, "xformers_attention": true } tokenizer_config.json: 0.00B [00:00, ?B/s] tokenizer_config.json: 141kB [00:00, 40.5MB/s] tokenizer.model: 0%| | 0.00/587k [00:00 [2025-10-30 00:38:10,859] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:279] [PID:3449] BOS: 1 / [2025-10-30 00:38:10,859] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:280] [PID:3449] PAD: 32768 / [2025-10-30 00:38:10,860] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:281] [PID:3449] UNK: 0 / [2025-10-30 00:38:10,860] [INFO] [axolotl.utils.data.shared.load_preprocessed_dataset:481] [PID:3449] Unable to find prepared dataset in last_run_prepared/7b29567c6088f93eb77efbc6c1f13463 [2025-10-30 00:38:10,860] [INFO] [axolotl.utils.data.sft._load_raw_datasets:320] [PID:3449] Loading raw datasets... [2025-10-30 00:38:10,860] [WARNING] [axolotl.utils.data.sft._load_raw_datasets:322] [PID:3449] Processing datasets during training can lead to VRAM instability. Please pre-process your dataset using `axolotl preprocess path/to/config.yml`. README.md: 0.00B [00:00, ?B/s] README.md: 21.6kB [00:00, 24.9MB/s] dataset_autotrain_curriculum.jsonl: 0%| | 0.00/21.7M [00:00", line 1, in top-level template code File "/root/miniconda3/envs/py3.11/lib/python3.11/site-packages/jinja2/sandbox.py", line 401, in call return __context.call(__obj, *args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/envs/py3.11/lib/python3.11/site-packages/transformers/utils/chat_template_utils.py", line 447, in raise_exception raise jinja2.exceptions.TemplateError(message) jinja2.exceptions.TemplateError: Conversation roles must alternate user/assistant/user/assistant/... """ The above exception was the direct cause of the following exception: Traceback (most recent call last): File "", line 198, in _run_module_as_main File "", line 88, in _run_code File "/workspace/axolotl/src/axolotl/cli/train.py", line 121, in fire.Fire(do_cli) File "/root/miniconda3/envs/py3.11/lib/python3.11/site-packages/fire/core.py", line 135, in Fire component_trace = _Fire(component, args, parsed_flag_args, context, name) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/envs/py3.11/lib/python3.11/site-packages/fire/core.py", line 468, in _Fire component, remaining_args = _CallAndUpdateTrace( ^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/envs/py3.11/lib/python3.11/site-packages/fire/core.py", line 684, in _CallAndUpdateTrace component = fn(*varargs, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/cli/train.py", line 88, in do_cli return do_train(parsed_cfg, parsed_cli_args) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/cli/train.py", line 43, in do_train dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/common/datasets.py", line 59, in load_datasets train_dataset, eval_dataset, total_num_steps, prompters = prepare_datasets( ^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/utils.py", line 50, in wrapper return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/sft.py", line 65, in prepare_datasets return _prepare_standard_dataset(cfg, tokenizer, processor) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/sft.py", line 98, in _prepare_standard_dataset train_dataset, eval_dataset, prompters = loader.load(_load_datasets) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/lock.py", line 38, in load result = load_fn() ^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/sft.py", line 77, in _load_datasets train_dataset, eval_dataset, prompters = _load_and_prepare_datasets( ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/sft.py", line 503, in _load_and_prepare_datasets dataset, prompters = _load_tokenized_prepared_datasets( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/sft.py", line 299, in _load_tokenized_prepared_datasets dataset, prompters = _load_raw_datasets( ^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/sft.py", line 331, in _load_raw_datasets dataset_wrapper, dataset_prompter = _load_and_process_single_dataset( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/sft.py", line 401, in _load_and_process_single_dataset dataset_wrapper, dataset_prompter = get_dataset_wrapper( ^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/wrappers.py", line 123, in get_dataset_wrapper return _handle_loaded_strategy(dataset_strategy, dataset, dataset_kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/utils/data/wrappers.py", line 223, in _handle_loaded_strategy dataset_wrapper = wrap_dataset_for_tokenized_prompt( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/datasets.py", line 87, in wrap_dataset_for_tokenized_prompt return TokenizedPromptDataset(prompt_tokenizer, dataset, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/datasets.py", line 40, in __init__ self.process(dataset).data, ^^^^^^^^^^^^^^^^^^^^^ File "/workspace/axolotl/src/axolotl/datasets.py", line 62, in process return dataset.map( ^^^^^^^^^^^^ File "/root/miniconda3/envs/py3.11/lib/python3.11/site-packages/datasets/arrow_dataset.py", line 560, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/envs/py3.11/lib/python3.11/site-packages/datasets/arrow_dataset.py", line 3309, in map for rank, done, content in iflatmap_unordered( File "/root/miniconda3/envs/py3.11/lib/python3.11/site-packages/datasets/utils/py_utils.py", line 626, in iflatmap_unordered [async_result.get(timeout=0.05) for async_result in async_results] File "/root/miniconda3/envs/py3.11/lib/python3.11/site-packages/datasets/utils/py_utils.py", line 626, in [async_result.get(timeout=0.05) for async_result in async_results] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/envs/py3.11/lib/python3.11/site-packages/multiprocess/pool.py", line 774, in get raise self._value jinja2.exceptions.TemplateError: Conversation roles must alternate user/assistant/user/assistant/...